-
Notifications
You must be signed in to change notification settings - Fork 73
How to infer from new videos ? #12
Comments
Hi @lecidhugo, I have updated this issue: #5 |
Hi @LuoweiZhou, |
Hi @lecidhugo, yes you're right. For the frame-wise features, please refer to this answer. Note that when extracting the region features, we uniformly sample 10 frames from each video segment while for frame-wise features, we sample the entire video at 2fps. Yes, if your end goal is inference/testing, you do not need to have any caption annotations. |
Thank you @LuoweiZhou, |
@lecidhugo The definition of video segments can be found here. You will see the start/end timestamp of each segment in the annotation file. For short videos, you can also directly feed them into the model. GVD captions each video (segment) independently. |
Hi @LuoweiZhou,
My question is two-fold:
1)
I downloaded the pre-trained models and I tried first to run the example you provided for inference. I got this error:
IOError: [Errno 2] No such file or directory: u'data/anet/rgb_motion_1d/K6Tm5xHkJ5c_resnet.npy'
Below is the point where the error arises:
Loading the model save/anet-unsup-0-0-0-run1/model-best.pth...
Finetune param: ctx2pool_grd.0.weight
Finetune param: ctx2pool_grd.0.bias
Finetune param: vis_embed.0.weight
I verified that the file is missing but I do not know how to get it ( I saw a similar issue but I could not proceed with the provided answer as it was unclear for me)
2) I am wondering how can I use your code in order to infer from my own videos. Can you please guide me ?
Thanks in advance
The text was updated successfully, but these errors were encountered: