-
Notifications
You must be signed in to change notification settings - Fork 502
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Run Zero123 under 13GB RAM #187
Comments
Hi, @generatorman. We are actively addressing this issue, and you can refer to this pull request for more details. You can also consider reducing the |
Thank you for the response. The PR you linked to seems related to VRAM usage - the issue I'm facing is with RAM. For example, running the following command quickly uses up 13GB of RAM and crashes, without using any VRAM at all.
So currently it's bottlenecked by RAM usage rather than VRAM usage. Is there any quick fix I could apply? |
I think loading the zero123 guidance model requires lots of RAM. To address this, you could consider modify |
@generatorman Honestly, you're going to need plenty of RAM and VRAM to run this kind of model. It's inevitable at this stage. Over time the efficiency of the code will probably improve, but for now, you need a good GPU and a powerful system. I recommend we close this issue for now. |
Any idea what the minimum model would be required? |
24GB is not enough, I run it on nvidia A10, it failed as OOM:
|
It seems 40GB VRAM is enough. I run it on A100 40G successfully.
|
@davideuler I wondered if the peak of the VRAM varied a lot when you ran the code. I am using A6000 48G VRAM and running into the OOM at step 400 or so. RuntimeError: /tmp/pip-req-build-zhms_45x/include/tiny-cuda-nn/gpu_memory.h:584 cuMemCreate(&m_handles.back(), n_bytes_to_allocate, &prop, 0) failed: CUDA_ERROR_OUT_OF_MEMORY
Epoch 0: | | 415/? [03:30<00:00, 1.97it/s, train/loss=15.70] |
Trying to run zero123 on Colab free tier fails because loading the model uses up all 12.7GB of RAM and crashes. Using some techniques to avoid loading the full model into RAM on the way to the GPU will unlock broader use of this exciting model.
The text was updated successfully, but these errors were encountered: