You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+82-2
Original file line number
Diff line number
Diff line change
@@ -249,6 +249,86 @@ ControlNet-LLLite, a novel method for ControlNet with SDXL, is added. See [docum
249
249
250
250
## Change History
251
251
252
+
### Jan 27, 2024 / 2024/1/27: v0.8.3
253
+
254
+
- Fixed a bug that the training crashes when `--fp8_base` is specified with `--save_state`. PR [#1079](https://github.com/kohya-ss/sd-scripts/pull/1079) Thanks to feffy380!
255
+
-`safetensors` is updated. Please see [Upgrade](#upgrade) and update the library.
256
+
- Fixed a bug that the training crashes when `network_multiplier` is specified with multi-GPU training. PR [#1084](https://github.com/kohya-ss/sd-scripts/pull/1084) Thanks to fireicewolf!
257
+
- Fixed a bug that the training crashes when training ControlNet-LLLite.
-[Experimental] The `--fp8_base` option is added to the training scripts for LoRA etc. The base model (U-Net, and Text Encoder when training modules for Text Encoder) can be trained with fp8. PR [#1057](https://github.com/kohya-ss/sd-scripts/pull/1057) Thanks to KohakuBlueleaf!
267
+
- Please specify `--fp8_base` in `train_network.py` or `sdxl_train_network.py`.
268
+
- PyTorch 2.1 or later is required.
269
+
- If you use xformers with PyTorch 2.1, please see [xformers repository](https://github.com/facebookresearch/xformers) and install the appropriate version according to your CUDA version.
270
+
- The sample image generation during training consumes a lot of memory. It is recommended to turn it off.
271
+
272
+
-[Experimental] The network multiplier can be specified for each dataset in the training scripts for LoRA etc.
273
+
- This is an experimental option and may be removed or changed in the future.
274
+
- For example, if you train with state A as `1.0` and state B as `-1.0`, you may be able to generate by switching between state A and B depending on the LoRA application rate.
275
+
- Also, if you prepare five states and train them as `0.2`, `0.4`, `0.6`, `0.8`, and `1.0`, you may be able to generate by switching the states smoothly depending on the application rate.
276
+
- Please specify `network_multiplier` in `[[datasets]]` in `.toml` file.
277
+
- Some options are added to `networks/extract_lora_from_models.py` to reduce the memory usage.
278
+
-`--load_precision` option can be used to specify the precision when loading the model. If the model is saved in fp16, you can reduce the memory usage by specifying `--load_precision fp16` without losing precision.
279
+
-`--load_original_model_to` option can be used to specify the device to load the original model. `--load_tuned_model_to` option can be used to specify the device to load the derived model. The default is `cpu` for both options, but you can specify `cuda` etc. You can reduce the memory usage by loading one of them to GPU. This option is available only for SDXL.
280
+
281
+
- The gradient synchronization in LoRA training with multi-GPU is improved. PR [#1064](https://github.com/kohya-ss/sd-scripts/pull/1064) Thanks to KohakuBlueleaf!
282
+
- The code for Intel IPEX support is improved. PR [#1060](https://github.com/kohya-ss/sd-scripts/pull/1060) Thanks to akx!
283
+
- Fixed a bug in multi-GPU Textual Inversion training.
284
+
285
+
- (実験的) LoRA等の学習スクリプトで、ベースモデル(U-Net、および Text Encoder のモジュール学習時は Text Encoder も)の重みを fp8 にして学習するオプションが追加されました。 PR [#1057](https://github.com/kohya-ss/sd-scripts/pull/1057) KohakuBlueleaf 氏に感謝します。
-`.toml` example for network multiplier / ネットワーク適用率の `.toml` の記述例
303
+
304
+
```toml
305
+
[general]
306
+
[[datasets]]
307
+
resolution = 512
308
+
batch_size = 8
309
+
network_multiplier = 1.0
310
+
311
+
... subset settings ...
312
+
313
+
[[datasets]]
314
+
resolution = 512
315
+
batch_size = 8
316
+
network_multiplier = -1.0
317
+
318
+
... subset settings ...
319
+
```
320
+
321
+
322
+
### Jan 17, 2024 / 2024/1/17: v0.8.1
323
+
324
+
- Fixed a bug that the VRAM usage without Text Encoder training is larger than before in training scripts for LoRA etc (`train_network.py`, `sdxl_train_network.py`).
325
+
- Text Encoders were not moved to CPU.
326
+
- Fixed typos. Thanks to akx! [PR #1053](https://github.com/kohya-ss/sd-scripts/pull/1053)
- Diffusers, Accelerate, Transformers and other related libraries have been updated. Please update the libraries with [Upgrade](#upgrade).
@@ -257,7 +337,7 @@ ControlNet-LLLite, a novel method for ControlNet with SDXL, is added. See [docum
257
337
- This feature works only on Linux or WSL.
258
338
- Please specify `--torch_compile` option in each training script.
259
339
- You can select the backend with `--dynamo_backend` option. The default is `"inductor"`. `inductor` or `eager` seems to work.
260
-
- Please use `--spda` option instead of `--xformers` option.
340
+
- Please use `--sdpa` option instead of `--xformers` option.
261
341
- PyTorch 2.1 or later is recommended.
262
342
- Please see [PR](https://github.com/kohya-ss/sd-scripts/pull/1024) for details.
263
343
- The session name for wandb can be specified with `--wandb_run_name` option. PR [#1032](https://github.com/kohya-ss/sd-scripts/pull/1032) Thanks to hopl1t!
@@ -270,7 +350,7 @@ ControlNet-LLLite, a novel method for ControlNet with SDXL, is added. See [docum
0 commit comments