• DocumentCode
    71210
  • Title

    Weakly Supervised Semantic Segmentation with a Multiscale Model

  • Author

    Wang, Shuhui ; Wang, Yannan

  • Author_Institution
    National Engineering Laboratory for Video Technology, Key Laboratory of Machine Perception (MoE), School of EECS, Peking University, 2728 Science Buildings, China
  • Volume
    22
  • Issue
    3
  • fYear
    2015
  • fDate
    Mar-15
  • Firstpage
    308
  • Lastpage
    312
  • Abstract
    This letter addresses the problem of weakly supervised semantic segmentation. Given training images with only image level annotations (i.e., tags) where the precise locations of tags are unknown, we simultaneously segment the images and assign tags to image regions. In contrast to previous work which segmented images at a specified scale, in this letter we propose a multiscale model for semantically segmenting images in different granularities and exploiting the long-range contextual information between adjacent scales. Then, to capture the geometric context of semantic labels, we augment the multiscale model by (i) the object spatial prior, e.g., “sky” has high probability on the top of an image, and (ii) the object spatial correlations, e.g., “car” always appears above “road”. Finally, we present an iterative top-down bottom-up method to learn the multiscale model by recovering the pixel labels of training images. Experiments on the benchmark MSRC21 and LMO datasets demonstrate the improved performance of our method over previous weakly supervised methods and even over some fully supervised methods.
  • Keywords
    Buildings; Correlation; Image edge detection; Image segmentation; Roads; Semantics; Training; Image semantic segmentation; multiscale; weakly supervised learning;
  • fLanguage
    English
  • Journal_Title
    Signal Processing Letters, IEEE
  • Publisher
    ieee
  • ISSN
    1070-9908
  • Type

    jour

  • DOI
    10.1109/LSP.2014.2358562
  • Filename
    6899593