Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

t2v timing #52

Open
BurhanUlTayyab opened this issue Jun 9, 2024 · 5 comments
Open

t2v timing #52

BurhanUlTayyab opened this issue Jun 9, 2024 · 5 comments

Comments

@BurhanUlTayyab
Copy link

Hi

I've implemented Lumina-T2V model and training it on Panda dataset. The paper mentions initial training takes 8 GPUs. I assume they are 8xA100 80GBs (which I'm using). May I know how long does it take (in terms of GPU hours)?

@gaopengpjlab
Copy link
Contributor

The paper claim 128 GPU is a must for T2V training.

@BurhanUlTayyab
Copy link
Author

IMG_6929

The first stage here refers 8 GPUs. I assume they are A100s, if not please tell me. Also tell me the GPU hours, how many GPU hours have been spent on training for both stage 1 and 2

@leonardodora
Copy link

Hi

I've implemented Lumina-T2V model and training it on Panda dataset. The paper mentions initial training takes 8 GPUs. I assume they are 8xA100 80GBs (which I'm using). May I know how long does it take (in terms of GPU hours)?

hi,do you have any plan to release the t2v codes?

@BurhanUlTayyab
Copy link
Author

Hi
I've implemented Lumina-T2V model and training it on Panda dataset. The paper mentions initial training takes 8 GPUs. I assume they are 8xA100 80GBs (which I'm using). May I know how long does it take (in terms of GPU hours)?

hi,do you have any plan to release the t2v codes?

Here's the code for Text to video model (https://drive.google.com/file/d/1jAtojjVmpzKuafUaFZjUT_HueAaAdve3/view?usp=sharing). But we have to warn you, that 8 GPU pretraining doesn't give good results at all. I asked them and as mentioned in the paper, they pretrain on 128 GPUs, so I don't know how they are doing pretraining on 8 GPUs. This is also the reason, they are not releasing T2V model.

@leonardodora
Copy link

Hi
I've implemented Lumina-T2V model and training it on Panda dataset. The paper mentions initial training takes 8 GPUs. I assume they are 8xA100 80GBs (which I'm using). May I know how long does it take (in terms of GPU hours)?

hi,do you have any plan to release the t2v codes?

Here's the code for Text to video model (https://drive.google.com/file/d/1jAtojjVmpzKuafUaFZjUT_HueAaAdve3/view?usp=sharing). But we have to warn you, that 8 GPU pretraining doesn't give good results at all. I asked them and as mentioned in the paper, they pretrain on 128 GPUs, so I don't know how they are doing pretraining on 8 GPUs. This is also the reason, they are not releasing T2V model.

Thanks for your code!
Maybe you could use t2i as pretrained model. 8 GPUs for t2v training from scratch is so challenging!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants