-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Full-parameter finetune后,生成视频的主体空间扭曲 #116
Comments
想知道您使用了多少数据进行微调,推荐使用100条相似的视频,以及, 您使用了默认配置吗,能提供一下loss的下降情况吗 |
感谢您的回复!
000000.mp44000次迭代: 000000.mp4期待您的回复~ |
Yes, for lora, lr 1e-4~1e-3 is OK. But for full-parameter fine-tune, lr 1e-5 is OK. |
Are there other factors besides the learning rate? Because the learning rate I am currently using is 1e-5, but as the training progresses, I will still observe a gradual decline in spatial ability. |
Is the prompt you use, "spider making a web", too different from your sft training data? And what is the total batch size? |
The total batch size is 24*2, and I'm using 100w dataset by changing dataset part. Next, waiting for more iterations, I test the training again. Thanks a lot! |
看起来4000步的结果也还比较正常,请问这里说的扭曲问题具体是指什么呢? |
一开始说的扭曲就是空间结构会有一些不合理。 |
我也在尝试finetune,所以空间结构不合理的问题是靠调小学习率 + 长时间训练解决的么? |
目前看是这样 |
Hey everyone! I have a few questions on finetuning that I would love if you could answer:
Thanks to everyone in advance! I might bother you with some more questions |
@a-r-r-o-w Have you got any answers? I'm also very curious about. |
Hey, yes I do! We worked together with Yuxuan from the CogVideoX team here: https://github.com/a-r-r-o-w/cogvideox-factory
Let me know if I can help you with anything else! |
Thanks a lot for replying! They are very helpful. I have one more question: so we don't need to use EMA model to train? |
I think there was a good recent paper that showed EMA is not particularly helpful for LoRA training, but the results with it are not too qualitatively different without it. It's really hard to see any benefits on small scale runs atleast (<10k steps in my tests), given the added memory requirement |
@a-r-r-o-w hi, Can you tell us more about which bug is causing the problem? |
@a-r-r-o-w hi, Can you tell us more about which bug is causing the problem? (The artifact issue was because of a bug in the Diffusers training scripts, which should have been addressed in cog-factory by now) |
您好!非常棒的开源repo~
最近尝试了Lora和full-parameter的finetune,均使用同样的50个video,微调500次迭代,其余setting没有修改
发现full-parameter的微调后,生成视频的主体会非常扭曲,lora的微调形式没有这种明显扭曲
下面是同样的prompt: spider making a web的结果:
full-parameter微调后
000000.mp4
lora微调后
000000.mp4
不知道导致这一问题的原因是什么?是微调时的lr太高的原因吗?
期待您的回复,感谢
The text was updated successfully, but these errors were encountered: