Templates for training #335
elismasilva
started this conversation in
General
Replies: 1 comment 3 replies
-
likely the most you can 'get away with' is:
even then, with all of the optimisations, SDXL training will consume more than 22G of VRAM without DeepSpeed. SDXL is a large model. if you aren't re-parameterising the thing (eg. going from epsilon to x-prediction) you are likely better off training a LoRA. try USE_DORA=true for LoRA training. if you're really needing to optimise full u-net training at the expense of correctness guarantees, OneTrainer claims to have support for full u-net training on a 3060 Ti. that's not something i'm interested in implementing here, as i honestly believe that's just not a good idea. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, do you have some template or sample configuration file for training SDXL ? i am starting training and testing hyperparameters, but i am trying to find one configuration that combine low vram consumption and velocity on training. I changed my bucket to 768x768 put grandient to 4 and batch size to 1 with 16 images. My scheduler is 'sine', what params impact more in memory consumption?
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions