From 30b7b1c7549fbc1277e3a9f5eed73a13f2f0c0ba Mon Sep 17 00:00:00 2001 From: Masahiro Hiramori Date: Wed, 25 Sep 2024 21:52:26 +0900 Subject: [PATCH] [CI] Upgrade unity image tag to `20240917-153130-9f281758` (#17410) * upgrade docker image to `20240917-153130-9f281758` * fix dynamo test case * building torch requires c++ 17 * temporary skip jax gpu tests due to XlaRuntimeError --- ci/jenkins/unity_jenkinsfile.groovy | 8 ++--- src/contrib/msc/plugin/torch_codegen.cc | 2 +- tests/python/relax/test_frontend_dynamo.py | 2 +- tests/python/relax/test_frontend_stablehlo.py | 36 ++++++++++++++++++- 4 files changed, 41 insertions(+), 7 deletions(-) diff --git a/ci/jenkins/unity_jenkinsfile.groovy b/ci/jenkins/unity_jenkinsfile.groovy index 9b4f0009e344..2a7a4fee3797 100755 --- a/ci/jenkins/unity_jenkinsfile.groovy +++ b/ci/jenkins/unity_jenkinsfile.groovy @@ -30,14 +30,14 @@ import org.jenkinsci.plugins.pipeline.modeldefinition.Utils // NOTE: these lines are scanned by docker/dev_common.sh. Please update the regex as needed. --> -ci_lint = 'tlcpack/ci-lint:20240105-165030-51bdaec6' -ci_gpu = 'tlcpack/ci-gpu:20240105-165030-51bdaec6' -ci_cpu = 'tlcpack/ci-cpu:20240105-165030-51bdaec6' +ci_lint = 'tlcpack/ci_lint:20240917-153130-9f281758' +ci_gpu = 'tlcpack/ci_gpu:20240917-153130-9f281758' +ci_cpu = 'tlcpack/ci_cpu:20240917-153130-9f281758' ci_wasm = 'tlcpack/ci-wasm:v0.72' ci_i386 = 'tlcpack/ci-i386:v0.75' ci_qemu = 'tlcpack/ci-qemu:v0.11' ci_arm = 'tlcpack/ci-arm:v0.08' -ci_hexagon = 'tlcpack/ci-hexagon:20240105-165030-51bdaec6' +ci_hexagon = 'tlcpack/ci_hexagon:20240917-153130-9f281758' // <--- End of regex-scanned config. // Parameters to allow overriding (in Jenkins UI), the images diff --git a/src/contrib/msc/plugin/torch_codegen.cc b/src/contrib/msc/plugin/torch_codegen.cc index 4b8c24f17bbb..75471d85db0d 100644 --- a/src/contrib/msc/plugin/torch_codegen.cc +++ b/src/contrib/msc/plugin/torch_codegen.cc @@ -219,7 +219,7 @@ void TorchPluginCodeGen::CodeGenCmake(const std::set& devices) { flags.Set("PLUGIN_SUPPORT_TORCH", ""); CodeGenPreCmake(devices, flags); stack_.line() - .line("set(CMAKE_CXX_STANDARD 14)") + .line("set(CMAKE_CXX_STANDARD 17)") .line("list(APPEND CMAKE_PREFIX_PATH \"" + config()->torch_prefix + "\")") .line("find_package(Torch REQUIRED)"); Array includes, libs; diff --git a/tests/python/relax/test_frontend_dynamo.py b/tests/python/relax/test_frontend_dynamo.py index 21e1d82d28b5..28215e2e6806 100644 --- a/tests/python/relax/test_frontend_dynamo.py +++ b/tests/python/relax/test_frontend_dynamo.py @@ -223,7 +223,7 @@ def subgraph_1( ) -> R.Tensor((10,), dtype="float32"): # block 0 with R.dataflow(): - lv5: R.Tensor((10,), dtype="float32") = R.multiply(inp_11, inp_01) + lv5: R.Tensor((10,), dtype="float32") = R.multiply(inp_01, inp_11) gv1: R.Tensor((10,), dtype="float32") = lv5 R.output(gv1) return gv1 diff --git a/tests/python/relax/test_frontend_stablehlo.py b/tests/python/relax/test_frontend_stablehlo.py index f2d0461dda77..667953ab73ec 100644 --- a/tests/python/relax/test_frontend_stablehlo.py +++ b/tests/python/relax/test_frontend_stablehlo.py @@ -196,6 +196,10 @@ def main( @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_unary(): import jax @@ -229,6 +233,10 @@ def _round(x): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_binary(): import jax @@ -250,6 +258,10 @@ def fn(x, y): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_const(): import jax @@ -260,6 +272,10 @@ def fn(x): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_maximum(): import jax import jax.numpy as jnp @@ -271,6 +287,10 @@ def fn(x, y): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_minimum(): import jax import jax.numpy as jnp @@ -282,6 +302,10 @@ def fn(x, y): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_reduce(): import jax import jax.numpy as jnp @@ -293,6 +317,10 @@ def fn(x): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_reduce_window(): import jax from flax import linen as nn @@ -304,6 +332,10 @@ def fn(x): @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) +# TODO(mshr-h): may be fixed by upgrading jax to >=0.4.33 def test_dot_general(): import jax @@ -314,8 +346,10 @@ def fn(x, y): check_correctness(jax.jit(fn), input_shapes) -@pytest.mark.skip() @tvm.testing.requires_gpu +@pytest.mark.skip( + reason="jaxlib.xla_extension.XlaRuntimeError: FAILED_PRECONDITION: DNN library initialization failed." +) # TODO(yongwww): fix flaky error of "invalid device ordinal" def test_conv(): import jax