diff --git a/.github/workflows/test-mlperf-inference-intel b/.github/workflows/test-mlperf-inference-intel.yml similarity index 85% rename from .github/workflows/test-mlperf-inference-intel rename to .github/workflows/test-mlperf-inference-intel.yml index b52a7ff708..2a8e452f33 100644 --- a/.github/workflows/test-mlperf-inference-intel +++ b/.github/workflows/test-mlperf-inference-intel.yml @@ -2,7 +2,7 @@ name: MLPerf Inference Intel implementations on: schedule: - - cron: "54 14 * * *" #to be adjusted + - cron: "29 16 * * *" #to be adjusted jobs: build_nvidia: @@ -16,9 +16,9 @@ jobs: steps: - name: Test MLPerf Inference Intel ${{ matrix.model }} run: | - if [ -f "gh_action/bin/deactivate" ]; then source gh_action/bin/deactivate; fi - python3 -m venv gh_action - source gh_action/bin/activate + if [ -f "gh_action_conda/bin/deactivate" ]; then source gh_action_conda/bin/deactivate; fi + python3 -m venv gh_action_conda + source gh_action_conda/bin/activate export CM_REPOS=$HOME/GH_CM pip install --upgrade cm4mlops cm run script --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r4.1-dev --preprocess_submission=yes --execution_mode=valid --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=IntelSPR.24c --implementation=intel --backend=pytorch --category=datacenter --division=open --scenario=Offline --docker_dt=yes --docker_it=no --docker_cm_repo=gateoverflow@cm4mlops --adr.compiler.tags=gcc --device=cpu --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean --docker --quiet diff --git a/script/app-mlperf-inference-amd/_cm.yaml b/script/app-mlperf-inference-amd/_cm.yaml index 089a1be1eb..3882380372 100644 --- a/script/app-mlperf-inference-amd/_cm.yaml +++ b/script/app-mlperf-inference-amd/_cm.yaml @@ -146,6 +146,12 @@ variations: CM_MLPERF_DEVICE: gpu CM_MLPERF_DEVICE_LIB_NAMESPEC: cudart + rocm: + group: device + env: + CM_MLPERF_DEVICE: rocm + CM_MLPERF_DEVICE_LIB_NAMESPEC: rocm + openshift: group: backend default: true @@ -161,6 +167,10 @@ variations: deps: - tags: get,generic-python-lib,_torch_cuda + pytorch,rocm: + deps: + - tags: get,generic-python-lib,_torch,_rocm + pytorch,cpu: deps: - tags: get,generic-python-lib,_torch diff --git a/script/process-mlperf-accuracy/customize.py b/script/process-mlperf-accuracy/customize.py index 4a6ba1634c..bb5d7a2865 100644 --- a/script/process-mlperf-accuracy/customize.py +++ b/script/process-mlperf-accuracy/customize.py @@ -103,6 +103,9 @@ def preprocess(i): else: extra_options += f""" --compliance-images-path '{os.path.join(result_dir, "images")}' """ + if env.get('CM_COCO2014_SAMPLE_ID_PATH','') != '': + extra_options += f" --ids-path '{env['CM_COCO2014_SAMPLE_ID_PATH']}' " + if env.get('CM_SDXL_ACCURACY_RUN_DEVICE', '') != '': extra_options += f" --device '{env['CM_SDXL_ACCURACY_RUN_DEVICE']}' "