Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About pretrain time #21

Open
1787648106 opened this issue Jul 4, 2023 · 5 comments
Open

About pretrain time #21

1787648106 opened this issue Jul 4, 2023 · 5 comments

Comments

@1787648106
Copy link

Thank you very much for your excellent work. May I ask, what is your experimental hardware configuration? For example, the model of the GPU; In addition, how long did you complete the pre training on imagenet1k?

@wwqq
Copy link
Collaborator

wwqq commented Jul 8, 2023

Hi @1787648106 , V100*8. Tiny:~1.5 days; Small:~1.8 days; Base:~2.2 days Large:~2.5days

@1787648106
Copy link
Author

thx for your response ~

@1787648106
Copy link
Author

1787648106 commented Nov 15, 2023

Hi @1787648106 , V100*8. Tiny:~1.5 days; Small:~1.8 days; Base:~2.2 days Large:~2.5days

hello~I pretrained the model on ImageNet1k, but the training time kept about ~5 days. A800 * 8, base model version, amp on. Are there some accelerating skills used ? Thanks again

@1787648106
Copy link
Author

Hi @1787648106 , V100*8. Tiny:~1.5 days; Small:~1.8 days; Base:~2.2 days Large:~2.5days

hello~I pretrained the model on ImageNet1k, but the training time kept about ~5 days. A800 * 8, base model version, amp on. Are there some accelerating skills used ? Thanks again

@wwqq hello, I'm so sorry to bother you again. About the accelerating in pretrain.

@wwqq
Copy link
Collaborator

wwqq commented Nov 25, 2023

@1787648106 Hello, try setting the batch size and learning rate to twice their original values.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants