-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
QKeras to QONNX converter redux #28
Conversation
…on against so far unsupported features. More tests.
Could you also update this with the latest |
tf2onnx does not support NumPy 1.24.1 (NVIDIA/TensorRT#2557), can we revert the dependency to NumPy 1.23.5 ? |
Note, some additional diffs got introduced when upgrading pre-commit hooks (basically new black style), do we want to downgrade to previous black? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM!
For making new users aware of any limitations, I would suggest to link the first post in this PR from README.md under the QKeras heading, unless you'd like to introduce a new section on the Sphinx docs? If this sounds good I can merge this PR and add that link to README. |
Sounds good, I can add a new section in the documentation to include the limitations. |
Merged in its current state but @selwyn96 please feel free to open a new PR for new documentation. |
This is the PR for the QKeras to QONNX converter that has fixed some of the issues in the previous PR draft #7.
Quick Overview:
This branch addresses the problem with the convolution layers and provides new tests for checking the stability of the converter. Even with the current fixes, there are still some issues with the converter. Some have workarounds that are outlined below:
Workaround: To use Quantized-Relu as separate QActivation layers.
Workaround: Use Quantized bits only at the output of Dense/Conv2D layers