-
Notifications
You must be signed in to change notification settings - Fork 388
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The result of swin-small backbone on ADE #5
Comments
ADE20K (and Cityscapes) dataset has very large variance, please try to train the model for 5 times, the median results should match the one we reported. |
Thanks for your reply, I will try to train the network again. Is it because I just use 4 GPUs while you train the network with 8 GPUs? |
As long as you used the same batch size, the number of GPU does not matter. |
Thanks for your reply. I have trained the network multiple times and the best result is 50.9. I find the variance is significantly large. |
Hi, @bowenc0221 , would you mind providing the variance for those methods and dataset? It is difficult to evaluate the method to compare the methods since the variance is so large. |
I don't have the variance for Mask2Former, but for the ADE20K and Cityscapes dataset we report the median of 3 runs. You can find the standard deviation (std) on the ADE20K dataset in our MaskFormer paper (https://arxiv.org/abs/2107.06278). Mask2Former should have a similar std. |
Thanks for your reply. |
Hi,
I run Mask2Former on ADE (maskformer2_swin_small_bs16_160k.yaml) with 4 16GB V-100 GPUs. However, I can only achieve 49.6%, which is much worse than the reported result (51.3%). Could you provide the log for me to analysize the result?
Thanks
The text was updated successfully, but these errors were encountered: