Files
Obsidian/Paper/CLIP/Open-Vocabulary Semantic Segmentation.md

1.2 KiB

Introduction

  • Two-stage approach
    • Method
      1. Generate class-agnostic mask proposal.
      2. Leverage pre-trained CLIP to perform open-vocabulary classification.
    • Assumption
      1. The model can generate classagnostic mask proposals.
      2. Pre-trained CLIP can transfer its classification performance to masked image proposals.
    • Examination
      1. Using ground-truth masks as region proposal.
      2. Feed masked images to a pre-trained CLIP for classification.
      3. Get mIoU of 20.1% on the ADE20K-150 dataset.
      4. Use MaskFormer(a mask proposal generator trained on COCO) as an region proposal generator.
      5. Select the region proposals with highest overlap with ground-truth masks.
      6. Assign the object label to this region.
      7. This model reach mIoU of 66.5%. (despite imperfect region proposal)
    • Conclusion Pre-trained CLIP not performed well over masked images, we hypothesize that CLIP trained on natural image which are not cropped or noised by segmentation masks.

Vocabularies

  1. ground-truth masks: refer to the manually annotated masks or pixel-level labels that are used to define the correct segmentation of objects in an image. Each pixel in the ground-truth mask is assigned a specific class label corresponding to the object or region it belongs to.