Replies: 3 comments 3 replies
-
I think you should be able to with some slight modifications to torchserve. Basically you just need to load the TRTorch runtime in your handler and also insure that the input size to your model is in the range specified at compile time. I tried out creating a mar file from a TRTorch compile TorchScript file from the torch-model-archiver tool and that seemed to work. Let us know if you try it and if you are successful or hit any issues, we'd love to hear about it |
Beta Was this translation helpful? Give feedback.
-
Thanks for the heads up. I'll close this question for now and will let you know the results. |
Beta Was this translation helpful? Give feedback.
-
Hi! I'm experiencing a weird problem with Torch_TENSORRT - it seems that within a local script environment, performing inference with my model and its correspective RT version consists of a 3x speed up in inference time, which is awesome. The problem is that I can't replicate this speed up within torchserve. Even though I load the runtime in my handler ( Is someone experiencing the same problem? |
Beta Was this translation helpful? Give feedback.
-
❓ Question
Is it possbile to use TRTorch with TorchServe?
If not what is the best way to deploy TRTorch programs?
What you have already tried
In the documentation, it is said I can continue to use programs via PyTorch API
I have converted all my models to TRTorch.
Environment
conda
,pip
,libtorch
, source):Additional context
Beta Was this translation helpful? Give feedback.
All reactions