You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Use a symmetric quantization with no clipping error to improve llama perplexity (#5163)
Summary:
Refer to pytorch/ao#805 for the details.
With this change, the perplexity of a llama model is improved 4% on wikitext.
Reviewed By: mergennachin, helunwencser
Differential Revision: D62342523
Pulled By: iseeyuan
0 commit comments