We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NVIDIA GPUでの推論をより早く軽くしたい
TensorRT 10.4 supports operators in the inclusive range of opset 9 to opset 20 https://github.com/onnx/onnx-tensorrt/blob/10.4-GA/docs/operators.md
The text was updated successfully, but these errors were encountered:
Googlefanさんが昨日話していたのですが、対応が難しいのとCUDAがあるので十分早いのではないかっていう話があります。ただ挑戦する価値はあると思います。
Sorry, something went wrong.
参考程度に4060tiでは1秒で25回短いテキストを推論できます ただもっと早くするロマンはありますよね...!
bertのほうはtrt対応できました。 本体はほかの人に託します...
No branches or pull requests
NVIDIA GPUでの推論をより早く軽くしたい
The text was updated successfully, but these errors were encountered: