DocumentCode
71210
Title
Weakly Supervised Semantic Segmentation with a Multiscale Model
Author
Wang, Shuhui ; Wang, Yannan
Author_Institution
National Engineering Laboratory for Video Technology, Key Laboratory of Machine Perception (MoE), School of EECS, Peking University, 2728 Science Buildings, China
Volume
22
Issue
3
fYear
2015
fDate
Mar-15
Firstpage
308
Lastpage
312
Abstract
This letter addresses the problem of weakly supervised semantic segmentation. Given training images with only image level annotations (i.e., tags) where the precise locations of tags are unknown, we simultaneously segment the images and assign tags to image regions. In contrast to previous work which segmented images at a specified scale, in this letter we propose a multiscale model for semantically segmenting images in different granularities and exploiting the long-range contextual information between adjacent scales. Then, to capture the geometric context of semantic labels, we augment the multiscale model by (i) the object spatial prior, e.g., “sky” has high probability on the top of an image, and (ii) the object spatial correlations, e.g., “car” always appears above “road”. Finally, we present an iterative top-down bottom-up method to learn the multiscale model by recovering the pixel labels of training images. Experiments on the benchmark MSRC21 and LMO datasets demonstrate the improved performance of our method over previous weakly supervised methods and even over some fully supervised methods.
Keywords
Buildings; Correlation; Image edge detection; Image segmentation; Roads; Semantics; Training; Image semantic segmentation; multiscale; weakly supervised learning;
fLanguage
English
Journal_Title
Signal Processing Letters, IEEE
Publisher
ieee
ISSN
1070-9908
Type
jour
DOI
10.1109/LSP.2014.2358562
Filename
6899593
Link To Document