We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
generate_quant.py脚本执行量化脚本后,效果很差,如图: 我的量化过程如下:
python tools/llama_quant.py /data/chat/models/chinese-v-13b-hf ptb --wbits 4 --groupsize 128 --save /data/chat/models/chinese-v-13b-hf/pyllama-4b.pt
python tools/generate_quant.py --model_path "/data/chat/models/chinese-v-13b-hf" --quant_path "/data/chat/models/chinese-v-13b-hf/pyllama-4b.pt" --wbits 4
是不是哪个环节出了错误
Originally posted by @greatewei in #46 (comment)
The text was updated successfully, but these errors were encountered:
正常现象,我们使用的是pyllama的方案,4bit量化比8bit差很多。之后会考虑GPTQ-for-LLaMa的方法
Sorry, something went wrong.
量化工具已经更新,效果请查看:https://github.com/Facico/Chinese-Vicuna/blob/master/tools/readme_zh.md
No branches or pull requests
generate_quant.py脚本执行量化脚本后,效果很差,如图:
我的量化过程如下:
python tools/llama_quant.py /data/chat/models/chinese-v-13b-hf ptb --wbits 4 --groupsize 128 --save /data/chat/models/chinese-v-13b-hf/pyllama-4b.pt
进行了模型量化,最终输出了 pyllama-4b.pt文件python tools/generate_quant.py --model_path "/data/chat/models/chinese-v-13b-hf" --quant_path "/data/chat/models/chinese-v-13b-hf/pyllama-4b.pt" --wbits 4
是不是哪个环节出了错误
Originally posted by @greatewei in #46 (comment)
The text was updated successfully, but these errors were encountered: