Inconsistent results between TensorRT and ONNX inference for ReduceMax operator #3467
Labels
internal-bug-tracked
Tracked internally, will be fixed in a future release.
triaged
Issue has been triaged by maintainers
Description
I am encountering an issue when using TensorRT to load an ONNX model that contains a
ReduceMax
operator. The inference results obtained from TensorRT are inconsistent with the results obtained from running the same model using ONNX runtime.The model structure is as below:
Environment
TensorRT Version: 8.6.1.6
NVIDIA GPU: NVIDIA GeForce MX330
NVIDIA Driver Version: 470.182.03
CUDA Version: 11.4
CUDNN Version: 8.9.5
Operating System: Ubuntu 18.04
Python Version (if applicable): 3.8
Relevant Files
Model link: https://github.com/hongliyu0716/onnx_model/blob/main/ReduceMax.onnx
Steps To Reproduce
The error message is as below:
The text was updated successfully, but these errors were encountered: