Co-attention CNNs for unsupervised object co-segmentation

Kuang Jui Hsu, Yen Yu Lin, Yung Yu Chuang

Research output: Chapter in Book/Report/Conference proceedingChapterpeer-review

66 Scopus citations

Abstract

Object co-segmentation aims to segment the common objects in images. This paper presents a CNN-based method that is unsupervised and end-to-end trainable to better solve this task. Our method is unsupervised in the sense that it does not require any training data in the form of object masks but merely a set of images jointly covering objects of a specific class. Our method comprises two collaborative CNN modules, a feature extractor and a co-attention map generator. The former module extracts the features of the estimated objects and backgrounds, and is derived based on the proposed co-attention loss which minimizes inter-image object discrepancy while maximizing intra-image figure-ground separation. The latter module is learned to generated co-attention maps by which the estimated figure-ground segmentation can better fit the former module. Besides, the co-attention loss, the mask loss is developed to retain the whole objects and remove noises. Experiments show that our method achieves superior results, even outperforming the state-of-the-art, supervised methods.
Original languageAmerican English
Title of host publicationIJCAI'18: Proceedings of the 27th International Joint Conference on Artificial Intelligence
PublisherInternational Joint Conferences on Artificial Intelligence
Pages748-756
Number of pages9
ISBN (Print)9780999241127
StatePublished - Jul 2018

Publication series

NameIJCAI International Joint Conference on Artificial Intelligence
Volume2018-July

Fingerprint

Dive into the research topics of 'Co-attention CNNs for unsupervised object co-segmentation'. Together they form a unique fingerprint.

Cite this