ExpPoint-MAE: Better interpretability and performance for self-supervised point cloud transformers

19 Jun 2023  ·  Ioannis Romanelis, Vlassis Fotis, Konstantinos Moustakas, Adrian Munteanu ·

In this paper we delve into the properties of transformers, attained through self-supervision, in the point cloud domain. Specifically, we evaluate the effectiveness of Masked Autoencoding as a pretraining scheme, and explore Momentum Contrast as an alternative. In our study we investigate the impact of data quantity on the learned features, and uncover similarities in the transformer's behavior across domains. Through comprehensive visualiations, we observe that the transformer learns to attend to semantically meaningful regions, indicating that pretraining leads to a better understanding of the underlying geometry. Moreover, we examine the finetuning process and its effect on the learned representations. Based on that, we devise an unfreezing strategy which consistently outperforms our baseline without introducing any other modifications to the model or the training pipeline, and achieve state-of-the-art results in the classification task among transformer models.

PDF Abstract

Results from the Paper


Ranked #13 on 3D Point Cloud Classification on ScanObjectNN (OBJ-ONLY (OA) metric, using extra training data)

     Get a GitHub badge
Task Dataset Model Metric Name Metric Value Global Rank Uses Extra
Training Data
Result Benchmark
3D Point Cloud Classification ModelNet40 ExpPoint-MAE Overall Accuracy 94.2 # 14
3D Point Cloud Classification ScanObjectNN ExpPoint-MAE OBJ-BG (OA) 90.88 # 15
OBJ-ONLY (OA) 90.02 # 13

Methods


No methods listed for this paper. Add relevant methods here