-
Notifications
You must be signed in to change notification settings - Fork 240
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reenable scale unification #2199
Conversation
I put this out of the draft state to trigger the CI runs first. |
Codecov ReportAttention:
Additional details and impacted files@@ Coverage Diff @@
## develop #2199 +/- ##
===========================================
- Coverage 90.82% 84.56% -6.27%
===========================================
Files 498 498
Lines 45485 45482 -3
===========================================
- Hits 41314 38464 -2850
- Misses 4171 7018 +2847
... and 54 files with indirect coverage changes
Flags with carried forward coverage won't be shown. Click here to find out more.
|
4d689b6
to
3b73a97
Compare
ONNX E2E 489 shows accuracy improvement for densenet12 (+0.77% acc.) and resnet50-v2-7 (+0.1% acc) and no difference for other models. TF E2E 490 shows no significant changes to the regular nightly run. PTQ build 240 shows accuracy degradation of 0.1% on timm/dpn68 and of 0.2% timm/visformer_small - both are for torch backend only, and for both the INT8 metric for the model is still within 99% FP32. |
519ba21
to
caf4000
Compare
caf4000
to
44390c9
Compare
39933c5
to
8c4ce45
Compare
8e0e9bf
to
2fe50c5
Compare
7bc7988
to
ddd930c
Compare
ddd930c
to
37ac299
Compare
@KodiaqQ PTQ build 285 shows a single regression vs build 286 on timm/visformer_small. Overall metric is still within 99% FP32. |
If the PTQ build is red, then we need to update the reference for the timm/visformer_small model as well. We can do it in this PR or follow-up, doesn't matter. But we should do it to keep builds green. |
Changes
Restored the original, pre-#1778 logic of scale unification and added missing tests for the logic. Added
BatchNorm
as a quantizable operation (asymmetric, per-channel) to the CPU HW config to handle cases like densenet where batch norm is the first operation in a branch.Reason for changes
Scales are currently not correctly unified in cases such as #2195.
Related tickets
N/A
Tests
tests/common/quantization/test_quantizer_setup.py
tests/**/quantization/test_unified_scales.py
Fixes: #2195