Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Too Slow training #31

Open
LaFeuilleMorte opened this issue Dec 20, 2023 · 8 comments
Open

Too Slow training #31

LaFeuilleMorte opened this issue Dec 20, 2023 · 8 comments

Comments

@LaFeuilleMorte
Copy link

Hi, Thanks for your great work and the open source code. I encountered too slow training when doing training with my RTX 3090 machine. And it will take 5~6 minutes to do 50 iterations (8000 in total). And the whole training would take like over 10 hours. That's way longer than what it is in the paper. Am I miss doing something?
image

@Anttwo
Copy link
Owner

Anttwo commented Dec 20, 2023

Hi LaFeuilleMorte,

Indeed, the training time seems very long, 50 iterations should be very short at the beginning of training (0.06 minutes) and get to 0.2 minutes max after starting the surface regularization.

I have several questions for you:

  1. Do you have the laptop or desktop version of the RTX 3090?
  2. How much memory does it have?
  3. How many Gaussians do you have in your initial Gaussian Splatting?

@yuedajiong
Copy link

我的屌丝配置计算机:
My poor-configuration computer:

(2016,personal GPU workstation)
GPU: Titan-XP 12G GPU-memory
CPU: 12 cores
MEM: 32G
DISK: SSD

the training speed is acceptable,15000 iterations, about dozens of minutes.

@LaFeuilleMorte
Copy link
Author

Hi, Thanks for the reply
Do you have the laptop or desktop version of the RTX 3090?
It's a desktop one.

How much memory does it have?
24GB

How many Gaussians do you have in your initial Gaussian Splatting?
image

@LaFeuilleMorte
Copy link
Author

我的屌丝配置计算机: My poor-configuration computer:

(2016,personal GPU workstation) GPU: Titan-XP 12G GPU-memory CPU: 12 cores MEM: 32G DISK: SSD

the training speed is acceptable,15000 iterations, about dozens of minutes.

To my best understanding of the code, it took too much time on the function "coarse_training_with_density_regularization".

@Sbector
Copy link

Sbector commented Dec 25, 2023

Happy holidays!

Same problem here.

I'm trying with a NVIDIA GeForce GTX 1650:
image
This is the information about my model:
image

Thanks this incredible work!

@DanielChaseButterfield
Copy link

I also have this very same issue.

However, I'm only using a GeForce RTX 2060, and it only has 14 GB of VRAM, so that might be my issue (as opposed to an issue with the repository)

@DanielChaseButterfield
Copy link

DanielChaseButterfield commented Mar 2, 2024

Looking into this issue a little more, I want to ask: @LaFeuilleMorte, what is your GPU utilization versus GPU memory usage?

When running my model, it seems that almost the entirety of the memory is used, but the GPU itself is doing almost no work at all. I theorize that this could be because the CPU isn't getting information to the GPU fast enough, and so the bottleneck is the CPU.

image

Looking at the code, it seems that the model is only trained a single image at a time (i.e. the batch size is 1). I wonder if this is why the GPU has nothing to do. I tried changing the following parameter to a larger number of images, but it seems that at some point during development, this value was fixed to 1, as I get the following error if I try to change it.
image

image

@yuedajiong
Copy link

it looks that the GS does not support batch.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants