-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG] 使用MiniCPM-o 2.6 int4模型时,chat.py报错 #728
Comments
self.model = AutoModel.from_pretrained(model_path, trust_remote_code=True).to(dtype=torch.bfloat16) |
同样的问题 |
应该是int4版本的model.safetensors文件中缺少metadata引起的 |
int4 model doesn't contain metadata |
It can NOT work, after installation of AutoGPTQ. The issue duplicated, as below.
Successfully installed auto_gptq |
If you want to using int4 version for webdemo, you should change the model initialization by model = AutoGPTQForCausalLM.from_quantized(
'openbmb/MiniCPM-o-2_6-int4',
torch_dtype=torch.bfloat16,
device="cuda:0",
trust_remote_code=True,
disable_exllama=True,
disable_exllamav2=True
) |
是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
该问题是否在FAQ中有解答? | Is there an existing answer for this in FAQ?
当前行为 | Current Behavior
我下载了MiniCPM-o 2.6 int4模型,并把路径写入到chat.py入口函数的model_path变量,运行代码得到了如下报错:
是我调用方式有问题吗?还是别的什么原因?如何解决这个问题?
期望行为 | Expected Behavior
能和模型正常对话
复现方法 | Steps To Reproduce
No response
运行环境 | Environment
备注 | Anything else?
No response
The text was updated successfully, but these errors were encountered: