Exploring Data Efficient 3D Scene Understanding
with Contrastive Scene Contexts

Ji Hou1      Benjamin Garahm2      Matthias Nießner1      Saining Xie2     

1Technical University of Munich       2Facebook AI Research

In this paper, we explore data-efficient learning for 3D point cloud. As a first step towards this direction, we propose Contrastive Scene Contexts, a 3D pre-training method that makes use of both point-level correspondences and spatial contexts in a scene. Our method achieves state-of-the-art results on a suite of benchmarks where training data or labels are scarce. Our study reveals that exhaustive labelling of 3D point clouds might be unnecessary; and remarkably, on ScanNet, even using 0.1% of point labels, we still achieve 89% (instance segmentation) and 96% (semantic segmentation) of the baseline performance that uses full annotations.


We introduce the benchmark for data-efficent 3d scene understanding of various mainstream tasks.

Challenge coming soon ... stay tuned.



Paper - ArXiv - pdf (abs) | GitHub (coming soon)

If you find our work useful, please consider citing it:
                title={Exploring Data-Efficient 3D Scene Understanding with Contrastive Scene Contexts}, 
                author={Hou, Ji and Graham, Benjamin and Nie{\ss}ner, Matthias and Saining Xie},
                journal={arXiv preprint arXiv:2012.09165},


If you have questions regarding our project, please contact us directly by email: ji.hou at tum.de