-
Notifications
You must be signed in to change notification settings - Fork 242
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training without Sequence Parallelism but VIDEO_SYNC_GROUP #162
Comments
Hi, we do not use the sequence parallel during training. The |
Hi @jy0205, thank you for the answer. Pyramid-Flow/dataset/dataset_cls.py Line 192 in e4b02ef
|
I think video_sync_group doesn't split same video latent, but accept same video latent without splitting.
|
|
Yes, you are right. The |
All the stages employ the uniform sampling. We will make the video token sequence length-balanced (let the token length sum to be fixed) |
Hi all,
the provided script
train_pyramid_flow.sh
does not set the flaguse_sequence_parallel
. In that case, what is the purpose of usingVIDEO_SYNC_GROUP=8
? Why we want all workers to use the same video?The text was updated successfully, but these errors were encountered: