-
Notifications
You must be signed in to change notification settings - Fork 7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Failures of test_quantized_classification_model[resnet50] #4683
Comments
@NicolasHug Thanks for the ping. Prior to #4597 the quantized models were not tested for their expected values like the unquantized ones. So it's the first time we observe issues related to platform differences etc. Like you said, it's worth investigating more to identify the source of flakiness. The quantized models are expected to be more flaky due to the reduced precision and the fact we are using uninitialised weights. One more reason to add the pre-trained weights on the CI cache. I propose to monitor the situation and if this becomes more problematic than the already flaky tests reported at #4506 we should look for a more immediate solution. Thoughts? Edit: |
None of the tests in #4506 have been reported as failing on the internal CI so far, and yet they've been around for longer, so this one seems a bit more serious I think. |
@NicolasHug I wonder if this is because these tests are disabled already on FBcode? @prabhat00155 could you clarify if you are looking into this already? |
I don't think so: the disabled tests would still show up as "broken" in https://www.internalfb.com/intern/tests/search?search_id=757311188293548 Taking a few of the ones from #4506 randomly, they're all green: |
Thanks for checking. I confirm that the tests fail because:
From the monitor it also seems that a couple of times the test passed, which is also weird because we fix the seed and the platform remains the same. So it's unclear if the additional randomness comes from hardware or other sources. At any case I agree that perhaps the easiest fix is to add the model in the |
This task T103498945 blamed D31649969(#4605) for the failures. However, when I ran the test under fbcode on my devvm I didn't see the failure. I haven't tried running the test on the external repo. |
Looks like
test_quantized_classification_model[resnet50]
is failing in some PRs like https://app.circleci.com/pipelines/github/pytorch/vision/11583/workflows/84517aa3-fa6b-4527-8ee6-8a09ae76199f/jobs/900594and it's also failing internally: https://www.internalfb.com/intern/tests/search?search_id=757311188293548
It looks like the failure is related to the new expected values checks introduced in #4597
I think an easy fix is just to add resnet50 to the
quantized_flaky_models
list, perhaps there's a better solution. It seems that the errors are consistently the same across executions, with the sameatol
andrtol
differences:So there might be a source of variability that we're not controlling?
cc @datumbox @pmeier
The text was updated successfully, but these errors were encountered: