DualFormer: A unified bidirectional sequence-to-sequence learning

Jen-Tzung Chien, Wei Hsiang Chang

Research output: Contribution to journalConference articlepeer-review

9 Scopus citations

Abstract

This paper presents a new dual domain mapping based on a unified bidirectional sequence-to-sequence (seq2seq) learning. Traditionally, dual learning in domain mapping was constructed with intrinsic connection where the conditional generative models in two directions were mutually leveraged and combined. The additional feedback from the other generation direction was used to regularize sequential learning in original direction of domain mapping. Domain matching between source sequence and target sequence was accordingly improved. However, the reconstructions for knowledge in two domains were ignored. The dual information based on separate models in two training directions was not sufficiently discovered. To cope with this weakness, this study proposes a closed-loop seq2seq learning where domain mapping and domain knowledge are jointly learned. In particular, a new feature-level dual learning is incorporated to build a dualformer where feature integration and feature reconstruction are further performed to bridge dual tasks. Experiments demonstrate the merit of the proposed dualformer for machine translation based on the multi-objective seq2seq learning.

Original languageEnglish
Pages (from-to)7718-7722
Number of pages5
JournalICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
Volume2021-June
DOIs
StatePublished - 6 Jun 2021
Event2021 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2021 - Virtual, Toronto, Canada
Duration: 6 Jun 202111 Jun 2021

Keywords

  • Domain mapping
  • Dual learning
  • Machine translation
  • Sequence-to-sequence learning
  • Transformer

Fingerprint

Dive into the research topics of 'DualFormer: A unified bidirectional sequence-to-sequence learning'. Together they form a unique fingerprint.

Cite this