Weakly supervised classification model for zero-shot semantic segmentation
2020
As one of the most fundamental tasks in computer vision, semantic segmentation assigns per pixel prediction of object categories. Training a robust model for semantic segmentation is challenging since pixel-level annotations are expensive to obtain. To alleviate the burden of annotations, the authors propose a weakly-supervised framework for zero-shot semantic segmentation, which can segment images having target classes without any pixel-level labelled instances. Under the assumption that the accessibility to image-level annotations of target classes does not violate the principle of zero pixel-level label in zero-shot semantic segmentation, we utilised image-level annotations to improve the proposed model's ability to extract pixel-level features. Furthermore, unlike existing zero-shot semantic segmentation methods, which use semantic embeddings as class embeddings to transfer knowledge from source classes to target classes, we use image-level features as their class embeddings to transfer knowledge since the distribution of pixel-level features is more similar to the distribution of image-level features rather than the distribution of semantic embeddings. Experimental results on the PASCAL-VOC data set under different data splits demonstrate that the proposed model achieves promising results.
- Correction
- Source
- Cite
- Save
- Machine Reading By IdeaReader
23
References
0
Citations
NaN
KQI