-
Notifications
You must be signed in to change notification settings - Fork 74
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is it possible to control scale and zero point of full integer quantized (INT8) tflite model during conversion? #709
Comments
Yes I think so that it is very much related to this issue: #269. I also have normalized the coordinates to the range [0,1] like confidence values are in the same range. It is absolutely possible that I just should take them to separate outputs and not using concat which might confuse Tensorflow Quantization. I was just wondering if the output scale and zero_point could be frozen before starting quantization. Freezing of activations and weights can be done with other quantization libraries like with tfmot and AIMET but not sure with TFLiteConverter. Other possibility is to change them afterwards. |
Just rewrite flatbuffer as python code with onnx2tf/onnx2tf/utils/common_functions.py Line 4541 in 2ecc03f
|
Okey I will have a look at it and tell soon how it worked :) |
If there is no activity within the next two days, this issue will be closed automatically. |
Issue Type
Feature Request
OS
Linux
onnx2tf version number
1.20.0
onnx version number
1.16.2
onnxruntime version number
1.19.2
onnxsim (onnx_simplifier) version number
tensorflow version number
2.17.0
Download URL for ONNX
https://github.com/ultralytics/ultralytics/blob/main/docs/en/models/yolov8.md
SiLUs replaced with ReLU. 320 resolution.
Parameter Replacement JSON
-
Description
The text was updated successfully, but these errors were encountered: