You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello,Thank you very much for your project and algorithm contributions! I have some questions regarding the process of fine-tuning and quantizing the llama model using the Qalora algorithm:
1.Do I need to quantize the pre-trained weights downloaded from Hugging Face using AutoGPT before training (considering the error encountered when directly loading the pre-trained weights: _FileNotFoundError: [Errno 2] No such file or directory: '/home/ud202481521/llm_model/llama-2-7b-hf/quantize_config.json'_)?
2.If quantization is required first, can Qalora continue with fine-tuning based on the quantized weights? When using AutoGPT to quantize the pre-trained weights and then training with Qalora based on those weights, I received the output message: _QuantLinear with cuda backend not support trainable mode yet, Switch to the pytorch backend_.
Additionally, I encountered the error: _RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn._
I might be misunderstanding how to use the algorithm properly. Your help would be greatly appreciated.
The text was updated successfully, but these errors were encountered:
Hello,Thank you very much for your project and algorithm contributions! I have some questions regarding the process of fine-tuning and quantizing the llama model using the Qalora algorithm:
1.Do I need to quantize the pre-trained weights downloaded from Hugging Face using AutoGPT before training (considering the error encountered when directly loading the pre-trained weights:
_FileNotFoundError: [Errno 2] No such file or directory: '/home/ud202481521/llm_model/llama-2-7b-hf/quantize_config.json'_
)?2.If quantization is required first, can Qalora continue with fine-tuning based on the quantized weights? When using AutoGPT to quantize the pre-trained weights and then training with Qalora based on those weights, I received the output message:
_QuantLinear with cuda backend not support trainable mode yet, Switch to the pytorch backend_
.Additionally, I encountered the error:
_RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn._
I might be misunderstanding how to use the algorithm properly. Your help would be greatly appreciated.
The text was updated successfully, but these errors were encountered: