Skip to content

Commit

Permalink
Update compression config for openlm-research/open_llama_3b_v2 (#860)
Browse files Browse the repository at this point in the history
* Remove compression with all_layers=True for openlm-research/open_llama_3b_v2

* Fix sym parameter

* Add AWQ
  • Loading branch information
nikita-savelyevv authored Aug 16, 2024
1 parent 46f8858 commit 1b7bd9f
Showing 1 changed file with 8 additions and 1 deletion.
9 changes: 8 additions & 1 deletion optimum/intel/openvino/configuration.py
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,14 @@ class OVQuantizationMethod(str, Enum):
"THUDM/chatglm2-6b": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.72},
"Qwen/Qwen-7B-Chat": {"bits": 4, "sym": True, "group_size": 128, "ratio": 0.6},
"openlm-research/open_llama_3b": {"bits": 4, "sym": False, "group_size": 64, "all_layers": True},
"openlm-research/open_llama_3b_v2": {"bits": 4, "sym": True, "group_size": 64, "all_layers": True},
"openlm-research/open_llama_3b_v2": {
"bits": 4,
"sym": False,
"group_size": 64,
"ratio": 1.0,
"dataset": "wikitext2",
"quant_method": OVQuantizationMethod.AWQ,
},
"tiiuae/falcon-7b-instruct": {"bits": 4, "sym": True, "group_size": 64, "all_layers": True},
"psmathur/orca_mini_3b": {
"bits": 4,
Expand Down

0 comments on commit 1b7bd9f

Please sign in to comment.