LLM 自适应 Shift-SmoothQuant #1894
Unanswered
egbertngu66
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
用自适应 Shift-SmoothQuant量化LLM,能达到W8A8(per-tensor)可用吗?https://github.com/PaddlePaddle/PaddleSlim/blob/develop/docs/zh_cn/tutorials/quant/advanced_quantization.md 表中的A8是per-tensor量化还是per-token量化?有相应的实验配置吗
Beta Was this translation helpful? Give feedback.
All reactions