Skip to content
/ DTS Public

Code for "Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection"

License

Notifications You must be signed in to change notification settings

WoodwindHu/DTS

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DTS: Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection

DTS is an unsupervised domain adaption method on 3D object detection, which is accepted on CVPR2023.

Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection
Qianjiang Hu, Daizong Liu, Wei Hu

Copyright (C) 2023 Qianjiang Hu, Daizong Liu, Wei Hu

License: MIT for academic use.

Contact: Wei Hu ([email protected])

Introduction

3D object detection from point clouds is crucial in safety-critical autonomous driving. Although many works have made great efforts and achieved significant progress on this task, most of them suffer from expensive annotation cost and poor transferability to unknown data due to the domain gap. Recently, few works attempt to tackle the domain gap in objects, but still fail to adapt to the gap of varying beam-densities between two domains, which is critical to mitigate the characteristic differences of the LiDAR collectors. To this end, we make the attempt to propose a density-insensitive domain adaption framework to address the density-induced domain gap. In particular, we first introduce Random Beam Re-Sampling (RBRS) to enhance the robustness of 3D detectors trained on the source domain to the varying beam-density. Then, we take this pre-trained detector as the backbone model, and feed the unlabeled target domain data into our newly designed task-specific teacher-student framework for predicting its high-quality pseudo labels. To further adapt the property of density-insensitive into the target domain, we feed the teacher and student branches with the same sample of different densities, and propose an Object Graph Alignment (OGA) module to construct two object-graphs between the two branches for enforcing the consistency in both the attribute and relation of cross-density objects. Experimental results on three widely adopted 3D object detection datasets demonstrate that our proposed domain adaption method outperforms the state-of-the-art methods, especially over varying-density data.

Model Zoo

Model Zoo

nuScenes -> KITTI TASK

Car@R40 download
SECOND-IoU 66.6 pretrained | model
PV-RCNN 71.8 pretrained | model
PointPillar 51.8 pretrained | model

Waymo -> KITTI TASK

Car@R40 download
SECOND-IoU 71.5 model
PVRCNN 68.1 model
PointPillar 50.2 model

Waymo -> nuScenes TASK

Car@R40 download
SECOND-IoU 23.0 model
PVRCNN 26.2 model
PointPillar 21.5 model

Installation

Please refer to INSTALL.md for the installation of OpenPCDet.

Usage

Please refer to GETTING_STARTED.md to learn more usage about this project.

Citation

@inproceedings{hu2023density,
  title={Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection},
  author={Hu, Qianjiang and Liu, Daizong and Hu, Wei},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year={2023}
}

About

Code for "Density-Insensitive Unsupervised Domain Adaption on 3D Object Detection"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published