MaskBooster: End-to-End Self-Training for Sparsely Supervised Instance Segmentation

Authors

  • Shida Zheng Hikvision Research Institute
  • Chenshu Chen Hikvision Research Institute
  • Xi Yang Hikvision Research Institute
  • Wenming Tan Hikvision Research Institute

DOI:

https://doi.org/10.1609/aaai.v37i3.25481

Keywords:

CV: Segmentation, ML: Semi-Supervised Learning

Abstract

The present paper introduces sparsely supervised instance segmentation, with the datasets being fully annotated bounding boxes and sparsely annotated masks. A direct solution to this task is self-training, which is not fully explored for instance segmentation yet. In this paper, we propose MaskBooster for sparsely supervised instance segmentation (SpSIS) with comprehensive usage of pseudo masks. MaskBooster is featured with (1) dynamic and progressive pseudo masks from an online updating teacher model, (2) refining binary pseudo masks with the help of bounding box prior, (3) learning inter-class prediction distribution via knowledge distillation for soft pseudo masks. As an end-to-end and universal self-training framework, MaskBooster can empower fully supervised algorithms and boost their segmentation performance on SpSIS. Abundant experiments are conducted on COCO and BDD100K datasets and validate the effectiveness of MaskBooster. Specifically, on different COCO protocols and BDD100K, we surpass sparsely supervised baseline by a large margin for both Mask RCNN and ShapeProp. MaskBooster on SpSIS also outperforms weakly and semi-supervised instance segmentation state-of-the-art on the datasets with similar annotation budgets.

Downloads

Published

2023-06-26

How to Cite

Zheng, S., Chen, C., Yang, X., & Tan, W. (2023). MaskBooster: End-to-End Self-Training for Sparsely Supervised Instance Segmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(3), 3696-3704. https://doi.org/10.1609/aaai.v37i3.25481

Issue

Section

AAAI Technical Track on Computer Vision III