Beyond the Label Itself: Latent Labels Enhance Semi-supervised Point Cloud Panoptic Segmentation
DOI:
https://doi.org/10.1609/aaai.v38i2.27887Keywords:
CV: 3D Computer Vision, CV: Multi-modal Vision, CV: SegmentationAbstract
As the exorbitant expense of labeling autopilot datasets and the growing trend of utilizing unlabeled data, semi-supervised segmentation on point clouds becomes increasingly imperative. Intuitively, finding out more ``unspoken words'' (i.e., latent instance information) beyond the label itself should be helpful to improve performance. In this paper, we discover two types of latent labels behind the displayed label embedded in LiDAR and image data. First, in the LiDAR Branch, we propose a novel augmentation, Cylinder-Mix, which is able to augment more yet reliable samples for training. Second, in the Image Branch, we propose the Instance Position-scale Learning (IPSL) Module to learn and fuse the information of instance position and scale, which is from a 2D pre-trained detector and a type of latent label obtained from 3D to 2D projection. Finally, the two latent labels are embedded into the multi-modal panoptic segmentation network. The ablation of the IPSL module demonstrates its robust adaptability, and the experiments evaluated on SemanticKITTI and nuScenes demonstrate that our model outperforms the state-of-the-art method, LaserMix.Downloads
Published
2024-03-24
How to Cite
Chen, Y., Tan, X., Zhang, Z., Qu, Y., & Xie, Y. (2024). Beyond the Label Itself: Latent Labels Enhance Semi-supervised Point Cloud Panoptic Segmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(2), 1245-1253. https://doi.org/10.1609/aaai.v38i2.27887
Issue
Section
AAAI Technical Track on Computer Vision I