From c11a2f5ee48ce02e72b1a0fa90772ed21c57b0e6 Mon Sep 17 00:00:00 2001 From: Yanjun Qi / Jane Date: Sat, 13 Apr 2024 13:36:13 -0400 Subject: [PATCH] Update S0-L25.md --- _contents/S0-L25.md | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/_contents/S0-L25.md b/_contents/S0-L25.md index 236a67c6..f3a24700 100755 --- a/_contents/S0-L25.md +++ b/_contents/S0-L25.md @@ -39,7 +39,11 @@ In this session, our readings cover: ## More readings - +### Must know tools for training/finetuning LLM's - +1. Torchtune - Build on top of Pytorch, for training and finetuning LLM's. Uses yaml based configs for easily running experiments. Github - https://lnkd.in/ghu6wx9r +2. axolotl - Built on top on Huggigface peft and transformer library, supports fine-tuning a large number for models like Mistral, LLama etc. Provides support for techniques like RLHF, DPO, LORA, qLORA etc. Github - https://lnkd.in/gYpisva9 +3. LitGPT - Build on nanoGPT and Megatron, support pre-training and fine-tuning, has examples like Starcoder, TinyLlama etc. Github - https://lnkd.in/gKisgXms +4. Maxtext - Jax based library for training LLM's on Google TPU's with configs for models like Gemma, Mistral and LLama2 etc. Github - https://lnkd.in/gjeHvZF4 ### Instruction Tuning for Large Language Models: A Survey + https://arxiv.org/abs/2308.10792