You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I try to load the model with transformers, small_model = AutoModelForCausalLM.from_pretrained(approx_model_name, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True)
but error occurs, OSError: Unable to load weights from pytorch checkpoint file for '/mnt/data3/lyk/models/tinyllama-1.1b/pytorch_model.bin' at '/mnt/data3/lyk/models/tinyllama-1.1b/pytorch_model.bin'. If you tried to load a PyTorch model from a TF 2.0 checkpoint, please set from_tf=True.
and when I set from_tf=True, another error occurs AttributeError: module transformers has no attribute TFLlamaForCausalLM
My package is torch2.1.0, transformers 4.39.3
The text was updated successfully, but these errors were encountered:
I encountered the same issue. It seems to be a bug in scripts/convert_lit_checkpoint.py. The model cannot be loaded due to UnicodeDecodeError (transformers 4.40.1).
Traceback (most recent call last):
File "/home/user/.conda/envs/py39pt23/lib/python3.9/site-packages/transformers/modeling_utils.py", line 542, in load_state_dict
if f.read(7) == "version":
File "/home/user/.conda/envs/py39pt23/lib/python3.9/codecs.py", line 322, in decode
(result, consumed) = self._buffer_decode(data, self.errors, final)
UnicodeDecodeError: 'utf-8' codec can't decode byte 0xb4 in position 64: invalid start byte
It works in transformers 4.35.0, so I load the model in this version and save it again using standard API, then the model can be loaded from future transformers versions. Note that I have safetensors installed, so the local model is saved as model.safetensors
I try to load the model with transformers,
small_model = AutoModelForCausalLM.from_pretrained(approx_model_name, torch_dtype=torch.float16, device_map="auto", trust_remote_code=True)
but error occurs,
OSError: Unable to load weights from pytorch checkpoint file for '/mnt/data3/lyk/models/tinyllama-1.1b/pytorch_model.bin' at '/mnt/data3/lyk/models/tinyllama-1.1b/pytorch_model.bin'. If you tried to load a PyTorch model from a TF 2.0 checkpoint, please set from_tf=True.
and when I set from_tf=True, another error occurs
AttributeError: module transformers has no attribute TFLlamaForCausalLM
My package is torch2.1.0, transformers 4.39.3
The text was updated successfully, but these errors were encountered: