This repository contains lecture materials on fine-tuning a clinical domain LLM for KoSAIM 2024.
- How to build a clinical domain Large Language Model (LLM)?
- (Large) Language Model
- How to build a (Language) Langauge Model?
- Building an instruction-following LLM in the clinical domain
- Asclepius (Gweon and Kim et al., ACL 2024 Findings)
- Hands-on Session: Fine-tuning a clinical domain LLM
- Environment Setup & Colab Practice
- LLM memory layout
- Parameter-Efficient Fine-Tuning (LoRA/QLoRA)
- https://github.com/huggingface/transformers
- https://github.com/huggingface/trl
- https://github.com/huggingface/accelerate
- https://github.com/huggingface/peft
- https://github.com/bitsandbytes-foundation/bitsandbytes
- https://github.com/starmpcc/Asclepius
- starmpcc/Asclepius-Synthetic-Clinical-Notes
BERT
Devlin, Jacob. "Bert: Pre-training of deep bidirectional transformers for language understanding." arXiv preprint arXiv:1810.04805 (2018).GPT
Radford, Alec, et al. "Improving Language Understanding by Generative Pre-Training."T5
Raffel, Colin, et al. "Exploring the limits of transfer learning with a unified text-to-text transformer." Journal of machine learning research 21.140 (2020): 1-67.- Liu, Pengfei, et al. "Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing." ACM Computing Surveys 55.9 (2023): 1-35.
GPT-3
Brown, Tom, et al. "Language models are few-shot learners." Advances in neural information processing systems 33 (2020): 1877-1901.FLAN
Wei, Jason, et al. "Finetuned language models are zero-shot learners." arXiv preprint arXiv:2109.01652 (2021).Llama
Touvron, Hugo, et al. "Llama: Open and efficient foundation language models." arXiv preprint arXiv:2302.13971 (2023).RLHF
Ouyang, Long, et al. "Training language models to follow instructions with human feedback." Advances in neural information processing systems 35 (2022): 27730-27744.ChatGPT
https://openai.com/index/chatgpt/Llama 2
Touvron, Hugo, et al. "Llama 2: Open foundation and fine-tuned chat models." arXiv preprint arXiv:2307.09288 (2023).Llama 3
Dubey, Abhimanyu, et al. "The Llama 3 Herd of Models." arXiv preprint arXiv:2407.21783 (2024).DPO
Rafailov, Rafael, et al. "Direct preference optimization: Your language model is secretly a reward model." Advances in Neural Information Processing Systems 36 (2024).Alpaca
Taori, Rohan, et al. "Alpaca: A strong, replicable instruction-following model." Stanford Center for Research on Foundation Models. https://crfm. stanford. edu/2023/03/13/alpaca. html 3.6 (2023): 7.Asclepius
Kweon, Sunjun, et al. "Publicly Shareable Clinical Large Language Model Built on Synthetic Clinical Notes." arXiv preprint arXiv:2309.00237 (2023).Lora
Hu, Edward J., et al. "Lora: Low-rank adaptation of large language models." arXiv preprint arXiv:2106.09685 (2021).QLora
Dettmers, Tim, et al. "Qlora: Efficient finetuning of quantized llms." Advances in Neural Information Processing Systems 36 (2024).- https://github.com/starmpcc/KAIA-LLM-FT-2024
- Seongsu Bae ([email protected])
- Sujeong Im ([email protected])