- [NIPS2022]Generating training data with language models: Towards zero-shot language understanding
- [EMNLP2022]Progen: Progressive zero-shot dataset generation via in-context feedback
- [NIPS2023]When less is more: Investigating data pruning for pretraining llms at scale
- [ICML2024]Curated llm: Synergy of llms and data curation for tabular augmentation in ultra low-data regimes
- [NAACL2024]from quantity to quality: Boosting llm performance with self-guided data selection for instruction tuning
- [ACL2024]Superfiltering: Weak-to-strong data filtering for fast instruction-tuning
- [COLM2024]Instruction Mining: When Data Mining Meets Large Language Model Finetuning
- [ICLR2024]Alpagasus:Training a better alpaca with fewer data
- [ICLR2024]# instag: Instruction tagging for analyzing supervised fine-tuning of large language models
- [ICLR2024]What makes good data for alignment? a comprehensive study of automatic data selection in instruction tuning
- [ICML2024]Less: Selecting influential data for targeted instruction tuning
- [ICLR2024]Self-alignment with instruction backtranslation
- [ACL2024]One shot learning as instruction data prospector for large language models.
- [EMNLP2024]Clustering and ranking:Diversity-preserved instruction selection through expert-aligned quality estimation
- [ACL2024]IterSelectTune: An Iterative Training Framework for Efficient Instruction-Tuning Data Selection
-
Notifications
You must be signed in to change notification settings - Fork 0
JasonCZH4/Paper-List-for-Instruction-Tuning-Data-Selection
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
About
No description, website, or topics provided.
Resources
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published