Skip to content

Commit

Permalink
Update ITREX version in ONNXRT WOQ example and fix bugs in hf models (#…
Browse files Browse the repository at this point in the history
…1333)

Signed-off-by: yuwenzho <[email protected]>
  • Loading branch information
yuwenzho authored and mengniwang95 committed Nov 20, 2023
1 parent f760298 commit 0ca51a1
Show file tree
Hide file tree
Showing 4 changed files with 6 additions and 2 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -483,6 +483,8 @@ def eval_func(model, *args):
if model_args.model_name_or_path == 'mrm8488/spanbert-finetuned-squadv1':
fp32_op_names = ['/bert/embeddings/word_embeddings/Gather',
'/bert/encoder/layer.[5-7|9]/output/dense/MatMul']
elif model_args.model_name_or_path == 'salti/bert-base-multilingual-cased-finetuned-squad':
fp32_op_names = ['/bert/encoder/layer.[4-5]/output/dense/MatMul']
elif model_args.model_name_or_path == 'distilbert-base-uncased-distilled-squad':
fp32_op_names = ['/distilbert/transformer/layer.[1-5]/ffn/lin[1-2]/MatMul']
elif model_args.model_name_or_path == 'deepset/roberta-large-squad2':
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
import onnxruntime as ort
from torch.nn.functional import pad
from torch.utils.data import DataLoader
from intel_extension_for_transformers.evaluation.lm_eval import evaluate
from intel_extension_for_transformers.llm.evaluation.lm_eval import evaluate
from optimum.onnxruntime import ORTModelForCausalLM
from transformers import LlamaConfig, LlamaTokenizer

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,8 @@ pip install -r requirements.txt
```
> Note: Validated ONNX Runtime [Version](/docs/source/installation_guide.md#validated-software-environment).
> Note: Weight-only quantization in Intel® Neural Compressor is still under development. We encourage you to use the `master` branch to access the latest features.
## 2. Prepare Model

Note that this README.md uses meta-llama/Llama-2-7b-hf as an example. There are other models available that can be used for weight-only quantization. The following table shows a few models' configurations:
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,7 +26,7 @@
import onnxruntime as ort
from torch.nn.functional import pad
from torch.utils.data import DataLoader
from intel_extension_for_transformers.evaluation.lm_eval import evaluate
from intel_extension_for_transformers.llm.evaluation.lm_eval import evaluate
from optimum.onnxruntime import ORTModelForCausalLM
from transformers import LlamaConfig, LlamaTokenizer

Expand Down

0 comments on commit 0ca51a1

Please sign in to comment.