TY - GEN
T1 - Semantic segmentation based on stacked discriminative autoencoders and context-constrained weakly supervised learning
AU - Yao, Xiwen
AU - Han, Junwei
AU - Cheng, Gong
AU - Guo, Lei
N1 - Publisher Copyright:
© 2015 ACM.
PY - 2015/10/13
Y1 - 2015/10/13
N2 - In this paper, we focus on tacking the problem of weakly supervised semantic segmentation. The aim is to predict the class label of image regions under weakly supervised settings, where training images are only provided with image-level labels indicating the classes they contain. The main difficulty of weakly supervised semantic segmentation arises from the complex diversity of visual classes and the lack of supervision information for learning a multi-classes classifier. To conquer the challenge, we propose a novel discriminative deep feature learning framework based on stacked autoencoders (SAE) by integrating pairwise constraints to serve as a discriminative term. Furthermore, to mine effective supervision information, global context about co-occurrence of visual classes as well as local context around each image region is exploited as constraints for training a multi-class classifier. Finally, the classifier training is formulated as an ultimate optimization problem, which can be solved efficiently by an alternate iterative optimization method. Comprehensive experiments on the MSRC 21 dataset demonstrate the superior performance compared with several state-of-The-Art weakly supervised image segmentation methods. Categories and Subject Descriptors I.4.6 [Image Processing and Computer Vision]: Segmentation -pixel classification; I.4.6 [Image Processing and Computer Vision]: Feature Measurement-feature representation; General Terms Algorithms, Experimentation, Performance. Keywords Semantic segmentation; Stacked autoencoders; Discriminative feature learning; Weakly supervised learning.
AB - In this paper, we focus on tacking the problem of weakly supervised semantic segmentation. The aim is to predict the class label of image regions under weakly supervised settings, where training images are only provided with image-level labels indicating the classes they contain. The main difficulty of weakly supervised semantic segmentation arises from the complex diversity of visual classes and the lack of supervision information for learning a multi-classes classifier. To conquer the challenge, we propose a novel discriminative deep feature learning framework based on stacked autoencoders (SAE) by integrating pairwise constraints to serve as a discriminative term. Furthermore, to mine effective supervision information, global context about co-occurrence of visual classes as well as local context around each image region is exploited as constraints for training a multi-class classifier. Finally, the classifier training is formulated as an ultimate optimization problem, which can be solved efficiently by an alternate iterative optimization method. Comprehensive experiments on the MSRC 21 dataset demonstrate the superior performance compared with several state-of-The-Art weakly supervised image segmentation methods. Categories and Subject Descriptors I.4.6 [Image Processing and Computer Vision]: Segmentation -pixel classification; I.4.6 [Image Processing and Computer Vision]: Feature Measurement-feature representation; General Terms Algorithms, Experimentation, Performance. Keywords Semantic segmentation; Stacked autoencoders; Discriminative feature learning; Weakly supervised learning.
UR - http://www.scopus.com/inward/record.url?scp=84962796451&partnerID=8YFLogxK
U2 - 10.1145/2733373.2806319
DO - 10.1145/2733373.2806319
M3 - 会议稿件
AN - SCOPUS:84962796451
T3 - MM 2015 - Proceedings of the 2015 ACM Multimedia Conference
SP - 1211
EP - 1214
BT - MM 2015 - Proceedings of the 2015 ACM Multimedia Conference
PB - Association for Computing Machinery, Inc
T2 - 23rd ACM International Conference on Multimedia, MM 2015
Y2 - 26 October 2015 through 30 October 2015
ER -