You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
And this how I trained the model:
(1) pretrained att2in2 for 25 epochs with the same settings (the same spatial feature of image, the same batch size, schedule sampling strategy from 0, the same learning rate decay, and so on), and I obtained comparable results with yours.
(2) then I trained it with scst for another 35 epochs. Learning rate was fixed to 5e-5. The cache for computing CIDEr is coco-train-idxs.
Compared with your result, the CIDEr is worse, but others metrics are better. The result bothers me a little bit, which makes me doubt about my experiment settings.
Is there anything trivial details I missed?
I wonder the schedule-sampling used in pretrained model will affect the exploration of the RL, but I have not had it a try. Any advice will be appreciated. Thanks a lot.
The text was updated successfully, but these errors were encountered:
miracle24
changed the title
The performance of att2in2 with scst bothers me.
Ask for advice for att2in2 under scst training
Apr 12, 2018
Hi. I trained att2in2 model with the default settings, but I got a lower score than https://github.com/ruotianluo/ImageCaptioning.pytorch/issues/10。
Here is my result:
Bleu_1: 0.796 Bleu_2: 0.622 Bleu_3: 0.471 Bleu_4: 0.351 ROUGE_L: 0.561 CIDEr: 1.118
and result in https://github.com/ruotianluo/ImageCaptioning.pytorch/issues/10:
Bleu_1: 0.777 Bleu_2: 0.613 Bleu_3: 0.465 Bleu_4: 0.347 ROUGE_L: 0.560 CIDEr: 1.156
And this how I trained the model:
(1) pretrained att2in2 for 25 epochs with the same settings (the same spatial feature of image, the same batch size, schedule sampling strategy from 0, the same learning rate decay, and so on), and I obtained comparable results with yours.
(2) then I trained it with scst for another 35 epochs. Learning rate was fixed to 5e-5. The cache for computing CIDEr is coco-train-idxs.
Compared with your result, the CIDEr is worse, but others metrics are better. The result bothers me a little bit, which makes me doubt about my experiment settings.
Is there anything trivial details I missed?
I wonder the schedule-sampling used in pretrained model will affect the exploration of the RL, but I have not had it a try. Any advice will be appreciated. Thanks a lot.
The text was updated successfully, but these errors were encountered: