PointContrast: Unsupervised Pre-training for 3D Point Cloud Understanding

07/21/2020
by   Saining Xie, et al.
11

Arguably one of the top success stories of deep learning is transfer learning. The finding that pre-training a network on a rich source set (eg., ImageNet) can help boost performance once fine-tuned on a usually much smaller target set, has been instrumental to many applications in language and vision. Yet, very little is known about its usefulness in 3D point cloud understanding. We see this as an opportunity considering the effort required for annotating data in 3D. In this work, we aim at facilitating research on 3D representation learning. Different from previous works, we focus on high-level scene understanding tasks. To this end, we select a suite of diverse datasets and tasks to measure the effect of unsupervised pre-training on a large source set of 3D scenes. Our findings are extremely encouraging: using a unified triplet of architecture, source dataset, and contrastive loss for pre-training, we achieve improvement over recent best results in segmentation and detection across 6 different benchmarks for indoor and outdoor, real and synthetic datasets – demonstrating that the learned representation can generalize across domains. Furthermore, the improvement was similar to supervised pre-training, suggesting that future efforts should favor scaling data collection over more detailed annotation. We hope these findings will encourage more research on unsupervised pretext task design for 3D deep learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/01/2023

AD-PT: Autonomous Driving Pre-Training with Large-scale Point Cloud Dataset

It is a long-term vision for Autonomous Driving (AD) community that the ...
research
08/17/2021

RandomRooms: Unsupervised Pre-training from Synthetic Shapes and Randomized Layouts for 3D Object Detection

3D point cloud understanding has made great progress in recent years. Ho...
research
11/21/2022

Self-Supervised Pre-training of 3D Point Cloud Networks with Image Data

Reducing the quantity of annotations required for supervised training is...
research
01/18/2023

Contrastive Learning for Self-Supervised Pre-Training of Point Cloud Segmentation Networks With Image Data

Reducing the quantity of annotations required for supervised training is...
research
01/12/2023

CLIP2Scene: Towards Label-efficient 3D Scene Understanding by CLIP

Contrastive language-image pre-training (CLIP) achieves promising result...
research
02/14/2022

COLA: COarse LAbel pre-training for 3D semantic segmentation of sparse LiDAR datasets

Transfer learning is a proven technique in 2D computer vision to leverag...
research
07/30/2022

Revisiting the Critical Factors of Augmentation-Invariant Representation Learning

We focus on better understanding the critical factors of augmentation-in...

Please sign up or login with your details

Forgot password? Click here to reset