You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for the good work. I am trying to reproduce the numbers reported in the paper (Table 1, MeaCap-TF). The paper mentions a CIDEr score of 42.5 on the training free variant. I use the command python inference.py --use_prompt --memory_id cc3m --img_path ./image_example --lm_model_path ./checkpoints/CBART_one_billion to find the MS-COCO captions and use pycocoeval package to find the language metrics. Here are the numbers that I got
which seems lower than the numbers in the paper. Can you point me to the evaluation code in the codebase? I am using pycocoeval and not sure if that is the reason for a lower performance. Or let me know if I am missing something.
Thanks
The text was updated successfully, but these errors were encountered:
thechargedneutron
changed the title
Reproducing the MeaCap-TF on MS COCO dataset
Reproducing MeaCap-TF results on MS COCO dataset
Oct 31, 2024
I tried --prompt_ensembling and I get similar performance as I reported above. Do you have the generations for this training-free variant and also, the evaluation code, if possible? Thanks!
Hi,
Thanks for the good work. I am trying to reproduce the numbers reported in the paper (Table 1, MeaCap-TF). The paper mentions a CIDEr score of 42.5 on the training free variant. I use the command
python inference.py --use_prompt --memory_id cc3m --img_path ./image_example --lm_model_path ./checkpoints/CBART_one_billion
to find the MS-COCO captions and use pycocoeval package to find the language metrics. Here are the numbers that I gotwhich seems lower than the numbers in the paper. Can you point me to the evaluation code in the codebase? I am using pycocoeval and not sure if that is the reason for a lower performance. Or let me know if I am missing something.
Thanks
The text was updated successfully, but these errors were encountered: