Replies: 1 comment
-
|
What is currently happening (cause of the problem) You get: This means: PyTorch has not compiled any CUDA kernels for sm_120 (Blackwell Architecture of the RTX 50 series) that it can run on your GPU. This isn't a fault of your installation alone — it's the way PyTorch binaries are built. As of now, many official PyTorch wheels for Windows do not yet fully support sm_120 — especially the stable releases that install WebUI by default. This means specifically: That's why a simple torch.randn(…, device="cuda") fails with the same error. So it's not a WebUI bug per se, but a compatibility issue between PyTorch + CUDA + RTX 50 series. Possible ways it can work
Many users report that the Nightly builds have better sm_120 support. pip uninstall -y torch torchvision torchaudio The nightly binaries can contain sm_120 kernels so that CUDA operations can run. Disadvantages: Open source issue tracker shows sm_120 support is still a work in progress. There is no fixed version, but support is gradually being integrated — especially in newer CUDA 12.8/12.9 binaries. Concrete steps for testing (recommended) Enable WebUI venv Completely remove all torch packages Install Nightly binaries (cu128) PyTorch GPU test: python - <<'EOF' Start WebUI If the test shows that torch.cuda.is_available() is True and sm_120 is detected, GPU inference should be possible. I hope I could help you... good luck. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I’m having trouble getting AUTOMATIC1111 Stable Diffusion WebUI to run on my GPU.
CPU mode works fine and can generate images, but any attempt to use CUDA on my RTX 5070 fails, including model loading.
I’m trying to figure out whether I’m using the wrong WebUI branch, or there is a DEV / experimental build that actually works on RTX 50-series (sm_120), or this is a current limitation of Windows PyTorch builds.
What works
What does NOT work
CUDA error: no kernel image is available for execution on the deviceThis happens even with:
I’ve confirmed this is not just an A1111 issue — even a basic test like:
torch.randn(16, device="cuda")fails with the same error.
About the “DEV mode” WebUI
I tried using a WebUI fork/branch that advertises DEV mode / RTX 50-series support, but:
If there is a working DEV branch, a fork that already supports sm_120, or a known-good commit / release
I’d really appreciate a link.
System Specs
GPU: NVIDIA RTX 5070 (12GB, sm_120)
CPU: AMD Ryzen 9 7900X (12-core)
OS: Windows 11
Driver: NVIDIA Studio Driver 591.74
Driver CUDA Capability: 13.1 (per nvidia-smi)
CUDA Toolkit: 12.8 installed
Python: 3.10.6
PyTorch tested: Stable cu128 and Nightly cu128 (torch 2.11 dev)
What I’m trying to find out
Is there a fully working WebUI version or fork that supports RTX 50-series on Windows?
Is this a known limitation of Windows PyTorch binaries for sm_120?
Are people with RTX 50-series currently using: a specific fork, WSL2/Linux, or DirectML instead?
If I’m missing any important system info, please let me know and I’ll add it.
Thanks in advance — I’m happy to test things and report back.
Beta Was this translation helpful? Give feedback.
All reactions