You Don’t Only Look Once: Constructing Spatial-Temporal Memory for Integrated 3D Object Detection and Tracking
You Don’t Only Look Once: Constructing Spatial-Temporal Memory for Integrated 3D Object Detection and Tracking
Jiaming Sun*, Yiming Xie*, Siyu Zhang, Linghao Chen, Guofeng Zhang, Hujun Bao, Xiaowei Zhou
ICCV 2021
Install Pytorch. The code is tested with Python 3.6, Pytorch v1.2.
Compile the CUDA layers for PointNet++, which we used in the backbone network:
./install.sh
Install the following Python dependencies (with pip install -r requirements
):
Download the pretrained weights and put it under
PROJECT_PATH/log_scannet
.
You can also use gdown to download it in command line:
mkdir log_scannet && cd log_scannet
gdown --id 1h3IBhC-oHGmUSEw1zT-AM94gccTqRpFa
Download and extract ScanNet by following the instructions provided at http://www.scan-net.org/.
Download oriented boxes annotation.
You can also use gdown to download it in command line:
gdown --id 1N-XdIsSpI7PrKgGzwjUk1ZEjykrHr2Z7
[Expected directory structure of ScanNet (click to expand)]
PROJECTROOT
└───scannet
│ └───oriented_boxes_annotation_train.pkl
│ └───oriented_boxes_annotation_val.pkl
│ └───scans
│ | └───scene0000_00
│ | └───depth
│ | │ │ 0.png
│ | │ │ 1.png
│ | │ │ ...
│ | │ ...
│ └───...
To train a VoteNet model on Scannet data (fused scan):
Single GPU:
export CUDA_VISIBLE_DEVICES=0;python train.py --dataset scannet --log_dir log_scannet --num_point 16384 --batch_size 8
Multiple GPUs:
export CUDA_VISIBLE_DEVICES=0,1;python -m torch.distributed.launch --nproc_per_node=2 train.py --dataset scannet --log_dir log_scannet --num_point 16384 --batch_size 8 --distributed
To test the trained model with its checkpoint:
python eval.py --dataset scannet --checkpoint_path log_scannet/104_checkpoint.tar --dump_dir eval_scannet --num_point 16384 --cluster_sampling seed_fps --use_3d_nms --use_cls_nms --per_class_proposal --batch_size 1 --faster_eval --feedback
Example results will be dumped in the eval_scannet
folder (or any other folder you specify). In default we evaluate with both [email protected] and [email protected] with 3D IoU on oriented boxes.
Thanks to Charles R. Qi for opening source of his excellent works VoteNet.
If you find this code useful for your research, please use the following BibTeX entry.
@article{sun2021udolo,
title={{You Don't Only Look Once}: Constructing Spatial-Temporal Memory for Integrated 3D Object Detection and Tracking},
author={Sun, Jiaming and Xie, Yiming and Zhang, Siyu and Chen, Linghao and Zhang, Guofeng and Bao, Hujun and Zhou, Xiaowei},
journal={{ICCV}},
year={2021}
}
This work is affiliated with ZJU-SenseTime Joint Lab of 3D Vision, and its intellectual property belongs to SenseTime Group Ltd.
Copyright SenseTime. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.