Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does mmdeploy supports onnx model quantization(onnx model with fp16 mode)?? #818

Closed
sanjaypavo opened this issue Jul 26, 2022 · 4 comments
Closed
Assignees

Comments

@sanjaypavo
Copy link
Contributor

Hi. I need to deploy my model(any object detection model) in onnx format in fp16 mode.Is it possible in mmdeploy?

Thanks in advance..

@tpoisonooo
Copy link
Collaborator

In theory, fp16 mode is not a kind of quantization, it is just convert fp32 value with bf16 format.

onnx format with fp16 precision not tested now.

mmdeploy using ppq to quantize ncnn int8. please check

@tpoisonooo tpoisonooo self-assigned this Jul 27, 2022
@tpoisonooo
Copy link
Collaborator

Looks like I have to write an English version of the quantization doc +_+

@tpoisonooo
Copy link
Collaborator

#842

@JIAOJINYU
Copy link

@tpoisonooo
I apologize for the sudden bother you.
I would like to ask you 2 questions.

  1. Does mmdeploy only support fp16 level quantization for onnxruntime at this moment?
  2. I currently would like to quantize rtmpose to int8. I try to quantize rtmpose to int8 using onnxruntime's static quantization. but the accuracy of the quantized model has been zero. Can I modify the pose-detection_onnxruntime-fp16_static.py to pose-detection_onnxruntime-int8_static.py by myself to realize the int8 quantization using mmdeploy?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants