Skip to content

Commit

Permalink
fix style
Browse files Browse the repository at this point in the history
  • Loading branch information
rnwang04 committed Nov 26, 2024
1 parent b740e4d commit 4cdbb99
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 2 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,8 @@ def update_names_of_IR_and_export_blob(model, model_name, dir, compile_blob=True
core.set_property("NPU", {"NPU_COMPILATION_MODE_PARAMS":
"compute-layers-with-higher-precision=Sqrt,Power,ReduceMean,Add"})
core.set_property("NPU", {"PERFORMANCE_HINT": "LATENCY"})
if npu_dpu_groups is not None and os.environ.get("IPEX_LLM_NPU_DISABLE_COMPILE_OPT", "0") != "1":
if (npu_dpu_groups is not None
and os.environ.get("IPEX_LLM_NPU_DISABLE_COMPILE_OPT", "0") != "1"):
core.set_property("NPU", {"NPU_DPU_GROUPS": str(npu_dpu_groups)})

model = core.read_model(xml_path)
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -472,7 +472,7 @@ def convert_llm_for_deploy(model: torch.nn.Module,
# llama3.2 1B & # llama3.2 3B
embedding_post = True
cos_sin_input = True
fused_layers = 2
fused_layers = 1
else:
# for Llama3-8B
fused_layers = 2
Expand Down

0 comments on commit 4cdbb99

Please sign in to comment.