Official implementation of the following paper:
Event-based Video Reconstruction Using Transformer by Wenming Weng, Yueyi Zhang, Zhiwei Xiong. In ICCV 2021.
HQF, MVSEC and IJRR datasets can be produced via the instructions in this repo. Note that MVSEC and IJRR are cut for better evaluation, of which the exact cut time can be found in the supplementary material.
The pretrained model, which can reproduce the quantitative results in the paper, will be released in this site
To reconstruct the intensity image using our ET-Net, E2VID, E2VID+, FireNet, FireNet+, one can run the file scripts/eval.py
. The paths to the pretrained model, dataset and output files should be specified. We provide example in scripts/eval.py
, one can look into this script for details.
If you find this work helpful, please consider citing our paper.
@InProceedings{Weng_2021_ICCV,
author = {Weng, Wenming and Zhang, Yueyi and Xiong, Zhiwei},
title = {Event-based Video Reconstruction Using Transformer},
booktitle = {Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV)},
year = {2021},
}
Reducing the Sim-to-Real Gap for Event Cameras, ECCV'20
High Speed and High Dynamic Range Video with an Event Camera, TPAMI'19
If you have any problem about the released code, please do not hesitate to contact me with email ([email protected]).