TriCI : Triple Cross-Intra Branch Contrastive Learning for Point Cloud Analysis

Whereas contrastive learning eliminates the need for labeled data, existing methods may suffer from inadequate features due to the conventional single shared encoder structure and struggle to fully harness the rich spectrum of 3D augmentations. In this paper, we propose TriCI, a self-supervised meth...

Ausführliche Beschreibung

Bibliographische Detailangaben
Veröffentlicht in:IEEE transactions on visualization and computer graphics. - 1996. - PP(2024) vom: 20. Aug.
1. Verfasser: Shao, Di (VerfasserIn)
Weitere Verfasser: Lu, Xuequan, Wang, Weijia, Liu, Xiao, Mian, Ajmal Saeed
Format: Online-Aufsatz
Sprache:English
Veröffentlicht: 2024
Zugriff auf das übergeordnete Werk:IEEE transactions on visualization and computer graphics
Schlagworte:Journal Article
LEADER 01000naa a22002652 4500
001 NLM376504935
003 DE-627
005 20240821233135.0
007 cr uuu---uuuuu
008 240821s2024 xx |||||o 00| ||eng c
024 7 |a 10.1109/TVCG.2024.3445962  |2 doi 
028 5 2 |a pubmed24n1508.xml 
035 |a (DE-627)NLM376504935 
035 |a (NLM)39163181 
040 |a DE-627  |b ger  |c DE-627  |e rakwb 
041 |a eng 
100 1 |a Shao, Di  |e verfasserin  |4 aut 
245 1 0 |a TriCI  |b Triple Cross-Intra Branch Contrastive Learning for Point Cloud Analysis 
264 1 |c 2024 
336 |a Text  |b txt  |2 rdacontent 
337 |a ƒaComputermedien  |b c  |2 rdamedia 
338 |a ƒa Online-Ressource  |b cr  |2 rdacarrier 
500 |a Date Revised 20.08.2024 
500 |a published: Print-Electronic 
500 |a Citation Status Publisher 
520 |a Whereas contrastive learning eliminates the need for labeled data, existing methods may suffer from inadequate features due to the conventional single shared encoder structure and struggle to fully harness the rich spectrum of 3D augmentations. In this paper, we propose TriCI, a self-supervised method that designs a triple-branch contrastive learning architecture. During contrastive pre-training, we generate three augmented versions of each input point cloud sample and pair each augmented sample with the original one, resulting in three unique positive pairs. We subsequently feed the pairs into three distinct encoders, each of which extracts features from its corresponding input positive pair. We design a novel cross-branch contrastive loss and use it along with the intra-branch contrastive loss to jointly train our network. The proposed cross-branch loss effectively aligns the output features from different perspectives for pre-training and facilitates their integration for downstream tasks, particularly in object-level scenarios. The intra-branch loss helps maximize the feature correspondences within positive pairs. Extensive experiments demonstrate the superiority of our TriCI in self-supervised learning, and show its strong ability in enhancing the performance of downstream object classification and part segmentation tasks. Interestingly, our TriCI achieves a 92.9% accuracy for linear SVM evaluation on ModelNet40, exceeding its closest competitor by 1.7% and even exceeding some supervised methods 
650 4 |a Journal Article 
700 1 |a Lu, Xuequan  |e verfasserin  |4 aut 
700 1 |a Wang, Weijia  |e verfasserin  |4 aut 
700 1 |a Liu, Xiao  |e verfasserin  |4 aut 
700 1 |a Mian, Ajmal Saeed  |e verfasserin  |4 aut 
773 0 8 |i Enthalten in  |t IEEE transactions on visualization and computer graphics  |d 1996  |g PP(2024) vom: 20. Aug.  |w (DE-627)NLM098269445  |x 1941-0506  |7 nnns 
773 1 8 |g volume:PP  |g year:2024  |g day:20  |g month:08 
856 4 0 |u http://dx.doi.org/10.1109/TVCG.2024.3445962  |3 Volltext 
912 |a GBV_USEFLAG_A 
912 |a SYSFLAG_A 
912 |a GBV_NLM 
912 |a GBV_ILN_350 
951 |a AR 
952 |d PP  |j 2024  |b 20  |c 08