-
Notifications
You must be signed in to change notification settings - Fork 2.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Docs][torch.compile] NNCF quantization/compression #27934
[Docs][torch.compile] NNCF quantization/compression #27934
Conversation
212a24a
to
a39e498
Compare
44f5476
to
193ec5c
Compare
TorchServe is a performant, flexible, and easy to use tool for serving PyTorch models in production. For more information on the details of TorchServe, | ||
you can refer to `TorchServe github repository. <https://github.com/pytorch/serve>`__. With OpenVINO ``torch.compile`` integration into TorchServe you can serve | ||
PyTorch models in production and accelerate them with OpenVINO on various Intel hardware. Detailed instructions on how to use OpenVINO with TorchServe are | ||
available in `TorchServe examples. <https://github.com/pytorch/serve/tree/master/examples/pt2/torch_compile_openvino>`__ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @daniil-lyakhov : I was thinking here, as we mention about general TorchServe examples..
NNCF unlocks the full potential of low-precision OpenVINO kernels due to quantizers placement designed specifically for the OpenVINO. | ||
Advanced algorithms like ``SmoothQuant`` or ``BiasCorrection`` allows furher metrics improvement minimizing the outputs discrepancy between the original and compressed models. | ||
For further details, please see the `documentation <https://docs.openvino.ai/2024/openvino-workflow/model-optimization.html>`__ | ||
and a `tutorial <https://github.com/daniil-lyakhov/nncf/tree/develop/examples/post_training_quantization/torch_fx/resnet18>`__. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think link is incorrect. it should be official main repo
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
My bad, fixed, thanks!
d500073
to
8546fe9
Compare
8546fe9
to
c921d4d
Compare
c15ea7d
to
a2e8e0b
Compare
8bd6753
to
7cb0446
Compare
7cb0446
to
1a8a556
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
port: #27934 Co-authored-by: Daniil Lyakhov <[email protected]>
https://openvino-doc.iotg.sclab.intel.com/nncf-docs/openvino-workflow/torch-compile.html#model-quantization-and-weights-compression
Details:
torch.compile
documentation