-
Notifications
You must be signed in to change notification settings - Fork 384
Description
Please help, getting the following error running on Windows 11
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 3050 Laptop GPU : native
Using pytorch cross attention
File "C:\Users\m\miniconda3\envs\MagicQuill\lib\site-packages\transformers\modeling_utils.py", line 3790, in from_pretrained
raise ValueError(
ValueError:
Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit
the quantized model. If you want to dispatch the model on the CPU or the disk while keeping
these modules in 32-bit, you need to set load_in_8bit_fp32_cpu_offload=True and pass a custom
device_map to from_pretrained. Check
https://huggingface.co/docs/transformers/main/en/main_classes/quantization#offload-between-cpu-and-gpu
for more details.