A Cross Branch Fusion-Based Contrastive Learning Framework for Point Cloud Self-supervised Learning

📅 2024-03-18
🏛️ International Conference on 3D Vision
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
Existing self-supervised contrastive learning methods for point clouds suffer from isolated multi-branch feature encoding, where branches operate independently until the loss layer, limiting inter-branch semantic interaction. Method: We propose PoCCA—a Point Cloud Cross-Branch Attention framework—that introduces cross-branch attention at early encoder stages to enable explicit feature interaction and fusion, departing from conventional unidirectional independent encoding. PoCCA jointly optimizes point cloud augmentation strategies, a dual-branch encoder, and a sub-branch attention module, requiring no additional labels or training data. Contribution/Results: On benchmarks including ModelNet40, PoCCA achieves state-of-the-art performance in downstream classification and segmentation tasks under self-supervised settings, demonstrating that early-stage cross-branch information exchange significantly enhances representation discriminability and generalizability.

Technology Category

Application Category

📝 Abstract
Contrastive learning is an essential method in self-supervised learning. It primarily employs a multi-branch strategy to compare latent representations obtained from different branches and train the encoder. In the case of multi-modal input, diverse modalities of the same object are fed into distinct branches. When using single-modal data, the same input undergoes various augmentations before being fed into different branches. However, all existing contrastive learning frameworks have so far only performed contrastive operations on the learned features at the final loss end, with no information exchange between different branches prior to this stage. In this paper, for point cloud unsupervised learning without the use of extra training data, we propose a Contrastive Cross-branch Attention-based framework for Point cloud data (termed PoCCA), to learn rich $3 D$ point cloud representations. By introducing sub-branches, PoCCA allows information exchange between different branches before the loss end. Experimental results demonstrate that in the case of using no extra training data, the representations learned with our self-supervised model achieve state-of-the-art performances when used for downstream tasks on point clouds.
Problem

Research questions and friction points this paper is trying to address.

Enables information exchange between branches before loss calculation
Improves self-supervised 3D point cloud representation learning
Achieves state-of-the-art performance without extra training data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cross-branch fusion for point cloud learning
Contrastive learning with sub-branches exchange
Attention-based framework without extra data
🔎 Similar Papers
No similar papers found.
Chengzhi Wu
Chengzhi Wu
Vision and Fusion Laboratory (IES), Karlsruhe Institue of Technology
computer vision
Q
Qianliang Huang
Institute for Anthropomatics and Robotics, Karlsruhe Institute of Technology, Germany
K
Kun Jin
Department of Automation, Tsinghua University, China
Julius Pfrommer
Julius Pfrommer
Head of Department, Fraunhofer IOSB
AutomationOptimizationMachine LearningIndustrie 4.0
J
Jürgen Beyerer
Institute for Anthropomatics and Robotics, Karlsruhe Institute of Technology, Germany; Fraunhofer Institute of Optronics, System Technologies and Image Exploitation IOSB, Germany