Image-to-image Translation via Contour-consistency Networks

Hsiang Ying Wang, Hsin Chun Lin, Chih Hsien Hsia*, Natnuntnita Siriphockpirom, Hsien I. Lin, Yung Yao Chen

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review


In this paper, a novel framework for image-to-image translation, in which contour-consistency networks are used to solve the problem of inconsistency between the contours of generated and original images, is proposed. The objective of this study was to address the lack of an adequate training set. At the generator end, the original map is sampled by an encoder to obtain the encoder feature map; the attention feature map is then obtained using the attention module. Using the attention feature map, the decoder can ascertain where more conversions are required. The mechanism at the discriminator end is similar to that at the generator end. The map is sampled through an encoder to obtain the encoder feature map and then converted into the attention feature map. Finally, the map is classified by the classifier as real or fake. Experimental results demonstrate the effectiveness of the proposed method.

Original languageEnglish
Pages (from-to)515-522
Number of pages8
JournalSensors and Materials
Issue number2
StatePublished - 2022


  • Attention feature map
  • Contour-consistency networks
  • Image-to-image translation
  • Inconsistency problem


Dive into the research topics of 'Image-to-image Translation via Contour-consistency Networks'. Together they form a unique fingerprint.

Cite this