**✅ SUCCESS: oobabooga now works with ROCm 7.2 on Windows + AMD GPUs!** #7375
Replies: 1 comment
-
|
Update: oobabooga now works with ROCm 7.2 on Windows!* I successfully got text-generation-webui running on AMD hardware with the latest ROCm release. Here's the current status: Working ConfigurationHardware:
Software:
Installation Steps
python -m venv rocm_env
rocm_env\Scripts\activate
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm_sdk_core-7.2.0.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm_sdk_devel-7.2.0.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm_sdk_libraries_custom-7.2.0.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm-7.2.0.dev0.tar.gz
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/torch-2.9.1+rocmsdk20260116-cp312-cp312-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/torchaudio-2.9.1+rocmsdk20260116-cp312-cp312-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/torchvision-0.24.1+rocmsdk20260116-cp312-cp312-win_amd64.whl
cd text-generation-webui
pip install -r requirements/full/requirements_amd.txt
pip uninstall torchao -y
python server.py --listen --apiKey Settings for Optimal Performance
PerformanceAchieving ~22-24 tokens/s on 7B models with proper GPU utilization using SDPA attention. Known Issues
Bottom LineROCm 7.2 + oobabooga works on Windows with AMD GPUs! Use Transformers loader with SDPA attention for best results.
I successfully got text-generation-webui running on AMD hardware with the latest ROCm release. Here's the current status: Working ConfigurationHardware:
Software:
Installation Steps
python -m venv rocm_env
rocm_env\Scripts\activate
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm_sdk_core-7.2.0.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm_sdk_devel-7.2.0.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm_sdk_libraries_custom-7.2.0.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/rocm-7.2.0.dev0.tar.gz
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/torch-2.9.1+rocmsdk20260116-cp312-cp312-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/torchaudio-2.9.1+rocmsdk20260116-cp312-cp312-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.2/torchvision-0.24.1+rocmsdk20260116-cp312-cp312-win_amd64.whl
cd text-generation-webui
pip install -r requirements/full/requirements_amd.txt
pip uninstall torchao -y
python server.py --listen --apiKey Settings for Optimal Performance
PerformanceAchieving ~22-24 tokens/s on 7B models with proper GPU utilization using SDPA attention. Known Issues
Bottom LineROCm 7.2 + oobabooga works on Windows with AMD GPUs! Use Transformers loader with SDPA attention for best results. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Update: Reported to AMD team as ROCm Issue #5871
Sharing my experience getting oobabooga running on AMD 7900 XTX with Windows. Got very close but hit an AMD bug. Documenting for others.
AMD Radeon RX 7900 XTX + Windows 11 + ROCm 7.1.1: Model Loads Successfully, Crashes on Text Generation
TL;DR: Model loads to GPU perfectly, crashes immediately on first generation attempt. Root cause: HIP runtime bug in amdhip64_7.dll. Awaiting AMD fix.
Hardware/Software Configuration
GPU: AMD Radeon RX 7900 XTX (24GB VRAM)
OS: Windows 11
Driver: AMD Software PyTorch Edition 25.20.01.17 (driver store version 32.0.22001.17002)
Python: 3.12.10
PyTorch: 2.9.0+rocmsdk20251116 (ROCm 7.1.1)
oobabooga: Latest main branch (January 2026)
Installation Steps & Reproduction
Download from python.org
During install, check "Add Python to PATH" and "Disable path length limit"
cmdcd D:
git clone https://github.com/oobabooga/text-generation-webui
cd text-generation-webui
cmdpython -m venv venv
venv\Scripts\activate.bat
cmdpip install -r requirements\full\requirements.txt
cmdpip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/rocm_sdk_core-0.1.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/rocm_sdk_devel-0.1.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/rocm_sdk_libraries_custom-0.1.dev0-py3-none-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/rocm-0.1.dev0.tar.gz
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/torch-2.9.0+rocmsdk20251116-cp312-cp312-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/torchaudio-2.9.0+rocmsdk20251116-cp312-cp312-win_amd64.whl
pip install --no-cache-dir https://repo.radeon.com/rocm/windows/rocm-rel-7.1.1/torchvision-0.24.0+rocmsdk20251116-cp312-cp312-win_amd64.whl
6. Verify PyTorch sees GPU
cmdpython -c "import torch; print(torch.cuda.is_available())"
python -c "import torch; print(torch.cuda.get_device_name(0))"
Expected output: True and AMD Radeon RX 7900 XTX
7. Start oobabooga
cmdpython server.py
8. Reproduce the crash
Open browser to http://127.0.0.1:7860/
Go to Model tab
In "Download model or LoRA", enter: mistralai/Mistral-7B-Instruct-v0.2
Click Download (wait for completion, ~15GB)
Select the model from dropdown
Model loader: Transformers
attn-implementation: eager (or sdpa, both crash)
Click Load → ✅ Succeeds ("Successfully loaded" message appears)
Go to Chat tab
Type any message (e.g., "Hello")
Click Generate → ❌ Python crashes silently
Check Windows Event Viewer for amdhip64_7.dll crash
What Works ✅
PyTorch correctly detects 7900 XTX
Model downloads successfully
Model loads to GPU successfully (completes in ~22 seconds)
Shows "Successfully loaded" message
What Fails ❌
Text generation crashes immediately on first GPU compute
Python process terminates with no console error
Windows Event Viewer shows amdhip64_7.dll access violation
Root Cause
Windows Event Viewer shows:
Beta Was this translation helpful? Give feedback.
All reactions