Skip to content

[ICLR 2024] FreeReg: Image-to-Point Cloud Registration Leveraging Pretrained Diffusion Models and Monocular Depth Estimators

Notifications You must be signed in to change notification settings

WHU-USI3DV/FreeReg

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

This is the official PyTorch implementation of the following publication:

FreeReg: Image-to-Point Cloud Registration Leveraging Pretrained Diffusion Models and Monocular Depth Estimators
Haiping Wang, Yuan Liu, Bing Wang, Yujing Sun, Zhen Dong, Wenping Wang, Bisheng Yang
ICLR 2024
Paper | Project-page | Video

🔭 Introduction

TL;DR: FreeReg extracts cross-modality features from pretrained diffusion models and monocular depth estimators for accurate zero-shot image-to-point cloud registration.

Network

Abstract: Matching cross-modality features between images and point clouds is a fundamental problem for image-to-point cloud registration. However, due to the modality difference between images and points, it is difficult to learn robust and discriminative cross-modality features by existing metric learning methods for feature matching. Instead of applying metric learning on cross-modality data, we propose to unify the modality between images and point clouds by pretrained large-scale models first, and then establish robust correspondence within the same modality. We show that the intermediate features, called diffusion features, extracted by depth-to-image diffusion models are semantically consistent between images and point clouds, which enables the building of coarse but robust cross-modality correspondences. We further extract geometric features on depth maps produced by the monocular depth estimator. By matching such geometric features, we significantly improve the accuracy of the coarse correspondences produced by diffusion features. Extensive experiments demonstrate that without any task-specific training, direct utilization of both features produces accurate image-to-point cloud registration. On three public indoor and outdoor benchmarks, the proposed method averagely achieves a 20.6 percent improvement in Inlier Ratio, a three-fold higher Inlier Number, and a 48.6 percent improvement in Registration Recall than existing state-of-the-arts.

🆕 News

💻 Requirements

The code has been tested on:

  • Ubuntu 20.04
  • CUDA 11.7
  • Python 3.8.5
  • Pytorch 1.12.1
  • GeForce RTX 4090.

🔧 Installation

  • First, create the conda environment:
    conda env create -f environment.yaml
    conda activate freereg
    
  • Second, fetch the latest MiDaS repo via torch hub for Zoe:
    import torch
    torch.hub.help("intel-isl/MiDaS", "DPT_BEiT_L_384", force_reload=True)  # Triggers fresh download of MiDaS repo
    
  • Third, install MinkowskiEngine for FCGF:
    conda install openblas-devel -c anaconda
    git clone https://github.com/NVIDIA/MinkowskiEngine.git
    cd MinkowskiEngine
    python setup.py install --blas_include_dirs=${CONDA_PREFIX}/include --blas=openblas
    

🚅 Pretrained model

*********** NOTE [2024.11.08] ************

If you just want to use FreeReg to register your own data or have a try, we suggest to use Free-FreeReg, which is much faster and easy-to-use.

If you are interested in the Diffusion Features and Full Power of FreeReg, Keep Going!


FreeReg does not need any training but utilizes pretrained models of existing projects:

  • Diffusion Feature related, download them and place them in tools/controlnet/models directory:

    • Stable Diffusion v1.5 [ckpt];
    • ControlNet(SDv1.5) conditioning on depth images [ckpt-ft]; [23/11/23 Update: We fine-tune ControlNet with real RGB-D data for better performances.]
  • Depth Estimation related, download them and place them in tools/zoe/models directory:

    • Zoe-depth-n [ckpt] for indoor datasets such as 3dmatch/scannet;
    • Zoe-depth-nk [ckpt] for outdoor datasets such as kitti;
  • Geometric Feature related, download them and place them in tools/fcgf/models directory:

🔦 Demo

You can try FreeReg with the following command:

python demo.py

Then, you should obtain a demo.png that looks like:

demo

💾 Dataset

The datasets are accessible in [Baidu-Disk](code:35h9) and Google Cloud:

Please place the data to ./data.

✏️ Test

To eval FreeReg on three benchmarks, you can use the following commands:

python run.py --dataset 3dmatch --type dg
python run.py --dataset scannet --type dg
python run.py --dataset kitti --type dg

you can replace --type dg that uses fused features and Kabsch solver with --type d for only using diffusion features and pnp solver or --type g for only using geometric features and Kabsch solver.

💡 Citation

If you find this repo helpful, please give us a 😍 star 😍. Please consider citing FreeReg if this program benefits your project

@article{wang2023freereg,
  title={FreeReg: Image-to-Point Cloud Registration Leveraging Pretrained Diffusion Models and Monocular Depth Estimators},
  author={Haiping Wang and Yuan Liu and Bing Wang and Yujing Sun and Zhen Dong and Wenping Wang and Bisheng Yang},
  journal={arXiv preprint arXiv:2310.03420},
  year={2023}
}

🔗 Related Projects

We sincerely thank the excellent projects:

  • ControlNet for Diffusion Feature extraction;
  • Zoe-Depth for metric depth estimation on rgb images;
  • FCGF for Geometric Feature extraction;
  • IDC-DC for depth completion.

Releases

No releases published

Packages

No packages published

Languages