Skip to content

Bug fixes#1195

Merged
danielhanchen merged 29 commits intomainfrom
nightly
Oct 26, 2024
Merged

Bug fixes#1195
danielhanchen merged 29 commits intomainfrom
nightly

Conversation

@danielhanchen
Copy link
Copy Markdown
Contributor

No description provided.

danielhanchen and others added 29 commits October 21, 2024 01:02
* Fix DPO, ORPO (#1177)

* Fix TRL

* Update mistral.py

* Patch processing_class

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Installation guide (#1165)

* chore: update chat_templates.py (#1166)

orginal -> original

* Disable Flex Attention

* Update tokenizer_utils.py

* Update _utils.py

* n_items

* Update cross_entropy_loss.py

* Fix DPO, ORPO

* Update _utils.py

---------

Co-authored-by: timothelaborie <97834767+timothelaborie@users.noreply.github.com>
Co-authored-by: Ikko Eltociear Ashimine <eltociear@gmail.com>

* Add warning for missing Unpack and KwargsForCausalLM in older Transformers versions

---------

Co-authored-by: Daniel Han <danielhanchen@gmail.com>
Co-authored-by: timothelaborie <97834767+timothelaborie@users.noreply.github.com>
Co-authored-by: Ikko Eltociear Ashimine <eltociear@gmail.com>
* Enhance rotary embedding handling in LlamaAttention and LongRopeRotaryEmbedding

* Typo

* Improve rotary embedding handling in LlamaAttention to prevent errors with short KV cache

* Update llama.py

* Update llama.py

---------

Co-authored-by: Daniel Han <danielhanchen@gmail.com>
@danielhanchen danielhanchen merged commit d76eda4 into main Oct 26, 2024
@WizKnight
Copy link
Copy Markdown

Hi @danielhanchen 🤗, I'm excited about the idea of adding float8 + QLoRA finetuning support via Torch AO into Unsloth !
I'd like to contribute and work on this. Do you have any specific points to consider before I start?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

6 participants