Skip to content

Latest commit

 

History

History
83 lines (58 loc) · 3.37 KB

README.md

File metadata and controls

83 lines (58 loc) · 3.37 KB

EGIInet: Explicitly Guided Information Interaction Network for Cross-modal Point Cloud Completion

This repository contains the official implementation for "Explicitly Guided Information Interaction Network for Cross-modal Point Cloud Completion" (ECCV 2024)

Explicitly Guided Information Interaction Network for Cross-modal Point Cloud Completion, ECCV 2024

Introduction

This work is accepted by ECCV 2024. In comparison with previous methods that relied on the global semantics of input images, EGIInet efficiently combines the information from two modalities by leveraging the geometric nature of the completion task. Specifically, we propose an explicitly guided information interaction strategy supported by modal alignment for point cloud completion. First, we unified the encoding process to promote modal alignment. Second, we propose a novel explicitly guided information interaction strategy that could help the network identify critical information within images, thus achieving better guidance for completion. Extensive experiments demonstrate the effectiveness of our framework, and we achieved a new state-of-the-art (+16% CD over XMFnet) in benchmark datasets despite using fewer parameters than the previous methods.

Get Started

Requirement

  • python >= 3.6
  • PyTorch >= 1.8.0
  • CUDA == 11.8
  • easydict
  • opencv-python
  • transform3d
  • h5py
  • timm
  • open3d
  • tensorboardX
  • ninja == 1.11.1
  • torch-scatter
  • einops

Install PointNet++ utils, Density-aware Chamfer Distance and Furthest Point Sampling.

cd /models/pointnet2_batch
python setup.py install

cd ../../metrics/CD/chamfer3D/
python setup.py install

cd ../../EMD/
python setup.py install

cd ../../utils/furthestPointSampling/
python setup.py install

The code has been trained and tested with Python 3.9, PyTorch 2.1.2 and CUDA 11.8 on Ubuntu 20.04 and RTX 3080Ti.

Dataset

Download the ShapeNet-ViPC dataset First, please download the ShapeNetViPC-Dataset (143GB, code: ar8l). Then run cat ShapeNetViPC-Dataset.tar.gz* | tar zx, you will get ShapeNetViPC-Dataset contains three floders: ShapeNetViPC-Partial, ShapeNetViPC-GT and ShapeNetViPC-View.

For each object, the dataset include partial point cloud (ShapeNetViPC-Patial), complete point cloud (ShapeNetViPC-GT) and corresponding images (ShapeNetViPC-View) from 24 different views. You can find the detail of 24 cameras view in /ShapeNetViPC-View/category/object_name/rendering/rendering_metadata.txt.

Use the code in utils/ViPCdataloader.py to load the dataset.

Pre-trained models

Download the pre-trained models at Google drive

Evaluation

# Specify the checkpoint path in config_vipc.py
__C.CONST.WEIGHTS = "path to your checkpoint"

python main.py --test

Training

python main.py

Acknowledgement

The repository is based on SVDFormer, some of the code is borrowed from:

Thanks for their opensourceing.

The point clouds are visualized with blender.

License

This project is open sourced under MIT license.