Skip to content

Latest commit

 

History

History
33 lines (29 loc) · 1.04 KB

Mixtral-8x7B-Instruct-v0.1-asym-recipe.md

File metadata and controls

33 lines (29 loc) · 1.04 KB

This recipe is outdated, we recommend using symmetric quantization. You can remove --asym from the command.

A sample command to generate an INT4 model.

auto-round \
--model  mistralai/Mixtral-8x7B-Instruct-v0.1 \
--device 0 \
--group_size 128 \
--bits 4 \
--iters 1000 \
--nsamples 512 \
--asym \
--format 'auto_gptq,auto_round' \
--output_dir "./tmp_autoround"

Install lm-eval-harness from source, and the git id f3b7917091afba325af3980a35d8a6dcba03dc3f is used

Metric BF16 INT4
Avg. 0.7000 0.6977
mmlu 0.6885 0.6824
lambada_openai 0.7718 0.7790
hellaswag 0.6767 0.6745
winogrande 0.7687 0.7719
piqa 0.8351 0.8335
truthfulqa_mc1 0.4969 0.4884
openbookqa 0.3680 0.3720
boolq 0.8850 0.8783
rte 0.7184 0.7004
arc_easy 0.8699 0.8712
arc_challenge 0.6220 0.6229