-
Notifications
You must be signed in to change notification settings - Fork 73
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inference Fails on 16GB GPU even in Low-VRAM Mode #32
Comments
Hmm that's weird. It ran fine (without low VRAM mode) on my 4080 mobile and that one has 12GB of VRAM. It should use ~10GB on memory without low VRAM mode and around 6GB with low VRAM mode. Are you sure that there's nothing else running on the GPU when you run the model? What does |
Yeah it's odd. It happens in both low VRAM mode and not. Here's nvidia-smi when I run without low VRAM and yes, it'll stay at ~10GB for some time +-----------------------------------------------------------------------------------------+ +-----------------------------------------------------------------------------------------+ But then all of a sudden it'll spike to 15GB+ and crash |
Can you try setting |
I'm running into issues trying to run.py on the demo files. Command:
python run.py demo_files/examples/fish.png --output-dir output/ --low-vram-mode
Output:
RuntimeError: CUDA driver error: out of memory
CUDA Available: True
CUDA Version: 12.4
PyTorch Version: 2.5.1+cu124
GPU 0: Quadro RTX 5000
GPU 0 Memory: 16.11 GB
I already tried rebuilding the venv twice now, but still not working. It works when I run it on my CPU but ideally, I'd like to get it running on my GPU. Thanks!
The text was updated successfully, but these errors were encountered: