diff --git a/training/LoRA_Llama-3.1-10epochs.ipynb b/training/QLoRA_Llama-3.1-10epochs.ipynb similarity index 99% rename from training/LoRA_Llama-3.1-10epochs.ipynb rename to training/QLoRA_Llama-3.1-10epochs.ipynb index b09b92d..488d4f3 100644 --- a/training/LoRA_Llama-3.1-10epochs.ipynb +++ b/training/QLoRA_Llama-3.1-10epochs.ipynb @@ -293,7 +293,7 @@ "source": [ "# Model and tokenizer names\n", "base_model_name = \"/data/HF-MODEL/huggingface-model/Meta-Llama-3.1-8B/\"\n", - "new_model_name = \"Llama-3.1-8B-lora\" #You can give your own name for fine tuned model\n", + "new_model_name = \"Llama-3.1-8B-qlora\" #You can give your own name for fine tuned model\n", "\n", "# Tokenizer\n", "#llama_tokenizer = AutoTokenizer.from_pretrained(base_model_name, trust_remote_code=True, use_fast=True)\n", @@ -797,7 +797,7 @@ "# Reload model in FP16 and merge it with LoRA weights\n", "\n", "#base_model_name = \"/data/HF-MODEL/huggingface-model/Meta-Llama-3.1-8B/\"\n", - "#new_model_name = \"Llama-3.1-8B-lora\" #You can give your own name for fine tuned model\n", + "#new_model_name = \"Llama-3.1-8B-qlora\" #You can give your own name for fine tuned model\n", "\n", "base_model = AutoModelForCausalLM.from_pretrained(base_model_name)\n", "from peft import LoraConfig, PeftModel\n",