Skip to content

yeyan00/pointcloud-sota

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

22 Commits
 
 
 
 
 
 
 
 

Repository files navigation

awesome-pointcloud-sota

Collect and summarize point cloud sota methods.

Available Tasks

Tasks

Examples

Classification


Segmentation


Object Detection

Panoptic Segmentation

Registration

Reconstruction

Multi-modal

pointcloud with language model

Change-Detection

pointcloud change detection

dataset

  • [ModelNet] ModelNet . [classification]
  • [scanobjectnn] The dataset contains ~15,000 objects that are categorized into 15 categories with 2902 unique object instances [classification]
  • [ScanNet] Richly-annotated 3D Reconstructions of Indoor Scenes. [classification segmentation]
  • [S3DIS] The Stanford Large-Scale 3D Indoor Spaces Dataset. [segmentation]
  • [npm3d] A Large-scale Mobile LiDAR Dataset for Semantic Segmentation of Urban Roadways [segmentation]
  • [KITTI-360] Corresponding to over 320k images and 100k laser scans in a driving distance of 73.7km. annotate both static and dynamic 3D scene elements with rough bounding primitives and transfer this information into the image domain, resulting in dense semantic & instance annotations for both 3D point clouds and 2D images. [segmentation]
  • [semantic3d] Large-Scale Point Cloud Classification Benchmark! a large labelled 3D point cloud data set of natural scenes with over 4 billion points in total. It also covers a range of diverse urban scenes. [segmentation]
  • SemanticKITTI Sequential Semantic Segmentation, 28 classes, for autonomous driving. All sequences of KITTI odometry labeled. [segmentation]
  • [ScribbleKITTI] Choose SemanticKITTI for its current wide use and established benchmark,ScribbleKITTI contains 189 million labeled points corresponding to only 8.06% of the total point count [segmentation]
  • [STPLS3D] a large-scale photogrammetry 3D point cloud dataset, termed Semantic Terrain Points Labeling - Synthetic 3D (STPLS3D), which is composed of high-quality, rich-annotated point clouds from real-world and synthetic environments.[segmentation]
  • [DALES] A Large-scale Aerial LiDAR Data Set for Point Cloud Segmentation,a new large-scale aerial LiDAR data set with nearly a half-billion points spanning 10 square kilometers of area [segmentation]
  • [SensatUrban] This dataset is an urban-scale photogrammetric point cloud dataset with nearly three billion richly annotated points, which is five times the number of labeled points than the existing largest point cloud dataset. Our dataset consists of large areas from two UK cities, covering about 6 km^2 of the city landscape. In the dataset, each 3D point is labeled as one of 13 semantic classes, such as ground, vegetation, car, etc.. [segmentation] *[H3D] H3D propose a benchmark consisting of highly dense LiDAR point clouds captured at four different epochs. The respective point clouds are manually labeled into 11 classes and are used to derive labeled textured 3D meshes as an alternative representation. UAV-based simultaneous data collection of both LiDAR data and imagery from the same platform,High density LiDAR data of 800 points/m² enriched by RGB colors of on board cameras incorporating a GSD of 2-3 cm [segmentation]
  • [KITTI] The KITTI Vision Benchmark Suite. [detection]
  • [Waymo] The Waymo Open Dataset is comprised of high resolution sensor data collected by Waymo self-driving cars in a wide variety of conditions.[detection segmentation]
  • [APOLLOSCAPE] The nuScenes dataset is a large-scale autonomous driving dataset.[detection segmentation]
  • [nuScenes] The nuScenes dataset is a large-scale autonomous driving dataset.[detection segmentation]
  • [3D Match] Keypoint Matching Benchmark, Geometric Registration Benchmark, RGB-D Reconstruction Datasets. [registration reconstruction ]
  • [ETH] Challenging data sets for point cloud registration algorithms [registration]
  • [objaverse] Objaverse-XL is an open dataset of over 10 million 3D objects! With it, we train Zero123-XL, a foundation model for 3D, observing incredible 3D generalization abilities.[Multi-modal]
  • [ScanRfer] 3D Object Localization in RGB-D Scans using Natural Language.[Multi-modal]
  • [DriveLM] DriveLM is an autonomous driving (AD) dataset incorporating linguistic information. Through DriveLM, we want to connect large language models and autonomous driving systems, and eventually introduce the reasoning ability of Large Language Models in autonomous driving (AD) to make decisions and ensure explainable planning. [Multi-modal]
  • [ScanQA] 3D Question Answering for Spatial Scene Understanding. A new 3D spatial understanding task for 3D question answering (3D-QA). In the 3D-QA task, models receive visual information from the entire 3D scene of a rich RGB-D indoor scan and answer given textual questions about the 3D scene [Multi-modal]
  • [urb3dcd-v2] The dataset is based on LoD2 models of the first and second districts of Lyon, France. To conduct fair qualitative and quantitative evaluation of point clouds change detection techniques. This first version of the dataset is composed of point clouds at a challenging low resolution of around 0.5 points/meter² [Change-Detection]

task sota

1. Classification

Model

Paper

Code

Year
PointView-GCN 3D shape classification with multi-view point clouds github 2021
PointGPT Auto-regressively Generative Pre-training from Point Clouds github 2023
point2vec Self-Supervised Representation Learning on Point Clouds github 2023
ULIP2 Towards Scalable Multimodal Pre-training for 3D Understanding github 2022
ReCon Contrast with Reconstruct: Contrastive 3D Representation Learning Guided by Generative Pretraining github 2023
RepSurf-U Surface Representation for Point Clouds github 2022
PointMLP Rethinking Network Design and Local Geometry in Point Cloud: A Simple Residual MLP Framework github 2022
PointNeXt PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies github 2022
PVT Point-Voxel Transformer for Point Cloud Learning github 2021
Point-MAE Masked Autoencoders for Point Cloud Self-supervised Learning github 2022
Point-M2AE Multi-scale Masked Autoencoders for Hierarchical Point Cloud Pre-training github 2022
Point-BERT Pre-Training 3D Point Cloud Transformers with Masked Point Modeling github 2021
PCT Point Cloud Transformer github 2020

2. Segmentation

Model

Paper

Code

Year
Swin3D++ Swin3D++: Effective Multi-Source Pretraining for 3D Indoor Scene Understanding github 2024
PointTransformerV3 Point Transformer V3: Simpler, Faster, Stronger github 2023
PonderV2 PonderV2: Pave the Way for 3D Foundation Model with A Universal Pre-training Paradigm github 2023
Swin3D A Pretrained Transformer Backbone for 3D Indoor Scene Understanding github 2023
Multi-dataset Point Prompt Training PPT:Towards Large-scale 3D Representation Learning with Multi-dataset Point Prompt Training github 2023
SphereFormer Spherical Transformer for LiDAR-based 3D Recognition github 2023
RangeFormer Rethinking Range View Representation for LiDAR Segmentation -- 2023
Window-Normalization Window Normalization: Enhancing Point Cloud Understanding by Unifying Inconsistent Point Densities github 2022
ptv2 Point Transformer V2: Grouped Vector Attention and Partition-based Pooling github 2022
stratified-transformer Stratified Transformer for 3D Point Cloud Segmentation github 2022
Superpoint Transformer Efficient 3D Semantic Segmentation with Superpoint Transformer github 2023
pointnext PointNeXt: Revisiting PointNet++ with Improved Training and Scaling Strategies github 2022
RepSurf Surface Representation for Point Clouds github 2022
CBL Contrastive Boundary Learning for Point Cloud Segmentation github 2022
FastPointTransformer Fast Point Transformer github 2022
PVKD Point-to-Voxel Knowledge Distillation for LiDAR Semantic Segmentation github 2022
Cylinder3D Cylindrical and Asymmetrical 3D Convolution Networks for LiDAR Segmentation github 2020
RandLA-Net RandLA-Net: Efficient Semantic Segmentation of Large-Scale Point Clouds github 2020
KPConv KPConv: Flexible and Deformable Convolution for Point Clouds github 2019

3. Detection

Model

Paper

Code

Year
GraphBEV Towards Robust BEV Feature Alignment for Multi-Modal 3D Object Detection -- 2024
DSVT Dynamic Sparse Voxel Transformer with Rotated Sets github 2023
BEVFusion an efficient and generic multi-task multi-sensor fusion framework github 2023
Pillar R-CNN Pillar R-CNN for Point Cloud 3D Object Detection github 2023
TransFusion Robust LiDAR-Camera Fusion for 3D Object Detection with Transformers github 2022
CenterFormer CenterFormer: Center-based Transformer for 3D Object Detection github 2022
CenterPoint Center-based 3D Object Detection and Tracking github 2020
TR3D Towards Real-Time Indoor 3D Object Detection github 2023
PV-RCNN++ Point-Voxel Feature Set Abstraction With Local Vector Representation for 3D Object Detection github 2021

4. Panoptic Segmentation

Model

Paper

Code

Year
P3Former Position-Guided Point Cloud Panoptic Segmentation Transformer github 2023
ISBNet a 3D Point Cloud Instance Segmentation Network with Instance-aware Sampling and Box-aware Dynamic Convolution github 2023
Mask3D Mask Transformer for 3D Instance Segmentation github 2022

5. registration

Model

Paper

Code

Year
DeformationPyramid Non-rigid Point Cloud Registration with Neural Deformation Pyramid github 2022
IMFNet Interpretable Multimodal Fusion for Point Cloud Registration github 2022
gedi Learning general and distinctive 3D local deep descriptors for point cloud registration github 2022
GeoTransformer Geometric Transformer for Fast and Robust Point Cloud Registration github 2022
D3Feat Joint Learning of Dense Detection and Description of 3D Local Features github 2020

6. reconstruction

Model

Paper

Code

Year
PoinTr PoinTr: Diverse Point Cloud Completion with Geometry-Aware Transformers github 2023
MaskSurf Masked Surfel Prediction for Self-Supervised Point Cloud Learning github 2022

7. multi-modal

Model

Paper

Code

Year
3D-LLM 3D-LLM: Injecting the 3D World into Large Language Models github 2023
PointLLM PointLLM: Empowering Large Language Models to Understand Point Clouds github 2023
CLIP-goes-3D CLIP goes 3D: Leveraging Prompt Tuning for Language Grounded 3D Recognition github 2023
LL3DA LL3DA: Visual Interactive Instruction Tuning for Omni-3D Understanding, Reasoning, and Planning github 2023
Chat-3D-v2 Chat-3D v2: Bridging 3D Scene and Large Language Models with Object Identifiers github 2024

8. change-detection

Model

Paper

Code

Year
DC3DCD unsupervised learning for multiclass 3D point cloud change detection github 2023
Siamese KPConv 3D multiple change detection from raw point clouds using deep learning github 2023
A Review Three Dimensional Change Detection Using Point Clouds: A Review github 2022

open libs

  • [Pointcept]
    Pointcept is a powerful and flexible codebase for point cloud perception research. (recommend)

  • [mmdetection3d]
    MMDetection3D is an open source object detection toolbox based on PyTorch, towards the next-generation platform for general 3D detection. It is a part of the OpenMMLab project developed by MMLab.

    • Support multi-modality/single-modality detectors out of box

    It directly supports multi-modality/single-modality detectors including MVXNet, VoteNet, PointPillars, etc.

    • Support indoor/outdoor 3D detection out of box

    It directly supports popular indoor and outdoor 3D detection datasets, including ScanNet, SUNRGB-D, Waymo, nuScenes, Lyft, and KITTI. For nuScenes dataset, we also support nuImages dataset.

    • Natural integration with 2D detection

    All the about 300+ models, methods of 40+ papers, and modules supported in MMDetection can be trained or used in this codebase.

    • High efficiency
  • [Robo3D]
    Robo3D is an evaluation suite heading toward robust and reliable 3D perception in autonomous driving. With it, we probe the robustness of 3D detectors and segmentors under out-of-distribution (OoD) scenarios against corruptions that occur in the real-world environment.

  • [open3d]
    Open3D is an open-source library that supports rapid development of software that deals with 3D data. The Open3D frontend exposes a set of carefully selected data structures and algorithms in both C++ and Python. The backend is highly optimized and is set up for parallelization. Open3D was developed from a clean slate with a small and carefully considered set of dependencies. It can be set up on different platforms and compiled from source with minimal effort. The code is clean, consistently styled, and maintained via a clear code review mechanism. Open3D has been used in a number of published research projects and is actively deployed in the cloud.

    Core features

    • Simple installation via conda and pip
    • 3D data structures
    • 3D data processing algorithms
    • Scene reconstruction
    • Surface alignment
    • PBR rendering
    • 3D visualization
    • Python binding
  • [OpenPCDet]
    OpenPCDet is a clear, simple, self-contained open source project for LiDAR-based 3D object detection. It is also the official code release of [PointRCNN], [Part-A2-Net], [PV-RCNN], [Voxel R-CNN], [PV-RCNN++] and [MPPNet].

  • [torch-points3d]

    Torch Points 3D is a framework for developing and testing common deep learning models to solve tasks related to unstructured 3D spatial data i.e. Point Clouds. The framework currently integrates some of the best published architectures and it integrates the most common public datasests for ease of reproducibility. It heavily relies on Pytorch Geometric and Facebook Hydra library thanks for the great work!

  • [learning3d]

    Learning3D is an open-source library that supports the development of deep learning algorithms that deal with 3D data. The Learning3D exposes a set of state of art deep neural networks in python. A modular code has been provided for further development. We welcome contributions from the open-source community.

  • [CloudCompare]

    CloudCompare is a 3D point cloud (and triangular mesh) processing software. It was originally designed to perform comparison between two 3D points clouds (such as the ones obtained with a laser scanner) or between a point cloud and a triangular mesh. It relies on an octree structure that is highly optimized for this particular use-case. It was also meant to deal with huge point clouds (typically more than 10 million points, and up to 120 million with 2 GB of memory).

About

Collect and summarize point cloud sota methods.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages