-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Performance on imagenet100 and imagenet1k #1
Comments
Hi, for imagenet100, we have got accuracy at 73+ (on par with that in CMC repo) with default config except for batch size = 128, which is the same with CMC repo. |
Could you share the config you use to get 73+? Thanks! |
You can just modify the batch size per gpu to 128/ngpu such as 16 if you use 8 gpu. |
BTW, I also tried batch size = 256, the accuracy is 70.540, which is also lower than the MoCo implementation in the CMC repo. |
I tried batchsize 128 and got result around 72.3, which is better than the previous but still slightly worse than your results. Just want to make sure I got everything right. Here're my commands:
And I'm running pytorch 1.4.0 and tochvision 0.5.0. I think the author mentioned that using alpha=0.99 is slightly better than 0.999. Do you notice the same thing? |
Thanks, I'll try these configs. Could you also share the configs to reproduce results on imagenet? |
Thanks! I reproduced 73+ results on imagenet100. Will try to run on imagenet1k. |
I will close this issue. If you have any questions, feel free to reopen it again. |
BTW, I got Acc@1 78.140% Acc@5 94.000% on imagenet100 with batch size 512, lr = 0.8, alpha = 0.99, K all. |
Thanks for sharing. Does “K all” mean K=len(dataset)?
…On Sat, Feb 29, 2020 at 7:26 PM 刘斌 ***@***.***> wrote:
BTW, I got ***@***.*** 78.140% ***@***.*** 94.000% on imagenet100 with batch size
512, lr = 0.8, alpha = 0.99, K all.
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub
<#1?email_source=notifications&email_token=AABNQNVPAEMV3SVVI6RB3YDRFHIVDA5CNFSM4KNKSRUKYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOENMTDYY#issuecomment-593048035>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AABNQNX3SO3ZRANAVYPTL3DRFHIVDANCNFSM4KNKSRUA>
.
|
Yes. |
I tried your large batch size settings and get Acc@1 around 76. Do you change anything for evaluation? How many GPUs do you use? |
Hi, sorry for the misleading message. Actually, the result of 78+ is obtained in my internal version, which uses the cosine learning rate decay with 5 epochs of warmup. |
Hi, I have updated the code and provide a script BTW, today I have merged a lot of updates from my internal version, such as warmup lr scheduler, add logger, support amp, etc. |
FYI, I have uploaded the checkpoint to onedrive which is pretrained on imagenet100 and achieve 78+. |
Is this a typo? This seems to be the dataset size of the Imagenet1k. |
Fixed. Thanks for your help. |
Have you tried similar large batch size settings on Imagenet1K? |
Actually, the key is the large base learning rate. For Imagenet1K, I also use the large batch size with a linear learning rate scale and warmup. But I have not tuned the base learning rate parameter. |
From my perspective, the imagenet100 is small, so the training is not sufficient. Then the large batch size and small alpha work well. |
@bl0 |
Have you tried your implementation on imagenet100 dataset? I'm getting accuracy at around 69.0 with default config (8 gpu, lr 0.03, bs 256), which is lower than the MoCo implementation in the CMC repo.
The text was updated successfully, but these errors were encountered: