Code and documentation to train Stanford's Alpaca models, and generate the data.
-
Updated
Mar 12, 2024 - Python
Code and documentation to train Stanford's Alpaca models, and generate the data.
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.
An Open-sourced Knowledgable Large Language Model Framework.
PhoGPT: Generative Pre-training for Vietnamese (2023)
A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.
Code for "Lion: Adversarial Distillation of Proprietary Large Language Models (EMNLP 2023)"
Finetune LLaMA-7B with Chinese instruction datasets
[ICLR 2024] Mol-Instructions: A Large-Scale Biomolecular Instruction Dataset for Large Language Models
[NeurIPS'23] "MagicBrush: A Manually Annotated Dataset for Instruction-Guided Image Editing".
BigCodeBench: The Next Generation of HumanEval
Code for "FollowBench: A Multi-level Fine-grained Constraints Following Benchmark for Large Language Models (ACL 2024)"
EditWorld: Simulating World Dynamics for Instruction-Following Image Editing
An benchmark for evaluating the capabilities of large vision-language models (LVLMs)
Instruction Following Agents with Multimodal Transforemrs
Is In-Context Learning Sufficient for Instruction Following in LLMs?
🌱 梦想家(DreamerGPT):中文大语言模型指令精调
EVE: Encoder-Free Vision-Language Models from BAAI
Code for the Paper "Grounding Hindsight Instructions in Multi-Goal Reinforcement Learning for Robotics"
A bachelor thesis on instruction following in MuJoCo.
Add a description, image, and links to the instruction-following topic page so that developers can more easily learn about it.
To associate your repository with the instruction-following topic, visit your repo's landing page and select "manage topics."