diff --git a/_contents/S0-L25.md b/_contents/S0-L25.md index 236a67c6..f3a24700 100755 --- a/_contents/S0-L25.md +++ b/_contents/S0-L25.md @@ -39,7 +39,11 @@ In this session, our readings cover: ## More readings - +### Must know tools for training/finetuning LLM's - +1. Torchtune - Build on top of Pytorch, for training and finetuning LLM's. Uses yaml based configs for easily running experiments. Github - https://lnkd.in/ghu6wx9r +2. axolotl - Built on top on Huggigface peft and transformer library, supports fine-tuning a large number for models like Mistral, LLama etc. Provides support for techniques like RLHF, DPO, LORA, qLORA etc. Github - https://lnkd.in/gYpisva9 +3. LitGPT - Build on nanoGPT and Megatron, support pre-training and fine-tuning, has examples like Starcoder, TinyLlama etc. Github - https://lnkd.in/gKisgXms +4. Maxtext - Jax based library for training LLM's on Google TPU's with configs for models like Gemma, Mistral and LLama2 etc. Github - https://lnkd.in/gjeHvZF4 ### Instruction Tuning for Large Language Models: A Survey + https://arxiv.org/abs/2308.10792