Is training FLUX with NF4 possible? #798
Replies: 6 comments 1 reply
-
we rely in third party libraries to do the quantisation because of how complex this gets while training. you will have to ask optimum-quanto devs that. also it already trains on consumer GPUs. |
Beta Was this translation helpful? Give feedback.
-
Hi, thank you for your reply. I would like to ask if it is possible to use |
Beta Was this translation helpful? Give feedback.
-
everything else the trainer depends on has to support it too. these are core dependencies like Accelerate and PEFT and the Diffusers projects. |
Beta Was this translation helpful? Give feedback.
-
I roughly understand what you mean. |
Beta Was this translation helpful? Give feedback.
-
yes that is inference related. trust me if it worked here sayak would open a pull request |
Beta Was this translation helpful? Give feedback.
-
@sayakpaul is indeed working on this :) |
Beta Was this translation helpful? Give feedback.
-
I saw on Reddit that someone achieved significant memory savings and speed improvements using the NF4 version of FLUX on Forge.
The quality impact wasn't very significant.
I wanted to ask if introducing NF4 in training is feasible.
Perhaps it could enable FLUX training on consumer-grade GPUs.
Beta Was this translation helpful? Give feedback.
All reactions