We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
在A100 80G上执行,按照要求,安装了相关的包后,按照“运行指南(int4-gptq篇)”执行第一步和第三步后,查看日志为: ,生成的文件如图: 执行第三步后,正常生成的文件是这几个吗? 最后执行python3 run.py --tokenizer_dir=Qwen-14B-Chat-Int4,发现预测结果不对,本次预测结果为: ,请问如何解决这个问题?
The text was updated successfully, but these errors were encountered:
补充执行第三步的指令为:python build.py --use_weight_only --weight_only_precision int4_gptq --per_group --hf_model_dir Qwen-14B-Chat-Int4 --quant_ckpt_path Qwen-14B-Chat-Int4
Sorry, something went wrong.
可以升级一下transformers版本试试。 顺便问问你用的哪个版本,是当前项目的main分支吗
用的是当前项目的main分支
升级transformers版本后就可以了,该问题是optimum和transformers版本不匹配导致的, 两者都用最新版就可以解决了。
No branches or pull requests
在A100 80G上执行,按照要求,安装了相关的包后,按照“运行指南(int4-gptq篇)”执行第一步和第三步后,查看日志为:
,生成的文件如图:
执行第三步后,正常生成的文件是这几个吗?
最后执行python3 run.py --tokenizer_dir=Qwen-14B-Chat-Int4,发现预测结果不对,本次预测结果为:
,请问如何解决这个问题?
The text was updated successfully, but these errors were encountered: