TY - GEN
T1 - Sequential learning and regularization in variational recurrent autoencoder
AU - Chien, Jen-Tzung
AU - Tsai, Chih Jung
N1 - Publisher Copyright:
© 2021 European Signal Processing Conference, EUSIPCO. All rights reserved.
Copyright:
Copyright 2021 Elsevier B.V., All rights reserved.
PY - 2021/1/24
Y1 - 2021/1/24
N2 - Latent variable model based on variational autoencoder (VAE) is influential in machine learning for signal processing. VAE basically suffers from the issue of posterior collapse in sequential learning procedure where the variational posterior easily collapses to a prior as standard Gaussian. Latent semantics are then neglected in optimization process. The recurrent decoder therefore generates noninformative or repeated sequence data. To capture sufficient latent semantics from sequence data, this study simultaneously fulfills an amortized regularization for encoder, extends a Gaussian mixture prior for latent variable, and runs a skip connection for decoder. The noise robust prior, learned from the amortized encoder, is likely aware of temporal features. A variational prior based on the amortized mixture density is formulated in implementation of variational recurrent autoencoder for sequence reconstruction and representation. Owing to skip connection, the sequence samples are continuously predicted in decoder with contextual precision at each time step. Experiments on language model and sentiment classification show that the proposed method mitigates the issue of posterior collapse and learns the meaningful latent features to improve the inference and generation for semantic representation.
AB - Latent variable model based on variational autoencoder (VAE) is influential in machine learning for signal processing. VAE basically suffers from the issue of posterior collapse in sequential learning procedure where the variational posterior easily collapses to a prior as standard Gaussian. Latent semantics are then neglected in optimization process. The recurrent decoder therefore generates noninformative or repeated sequence data. To capture sufficient latent semantics from sequence data, this study simultaneously fulfills an amortized regularization for encoder, extends a Gaussian mixture prior for latent variable, and runs a skip connection for decoder. The noise robust prior, learned from the amortized encoder, is likely aware of temporal features. A variational prior based on the amortized mixture density is formulated in implementation of variational recurrent autoencoder for sequence reconstruction and representation. Owing to skip connection, the sequence samples are continuously predicted in decoder with contextual precision at each time step. Experiments on language model and sentiment classification show that the proposed method mitigates the issue of posterior collapse and learns the meaningful latent features to improve the inference and generation for semantic representation.
KW - Bayesian learning
KW - Language model
KW - Recurrent neural network
KW - Sequential learning
KW - Variational autoencoder
UR - http://www.scopus.com/inward/record.url?scp=85099307314&partnerID=8YFLogxK
U2 - 10.23919/Eusipco47968.2020.9287616
DO - 10.23919/Eusipco47968.2020.9287616
M3 - Conference contribution
AN - SCOPUS:85099307314
T3 - European Signal Processing Conference
SP - 1613
EP - 1617
BT - 28th European Signal Processing Conference, EUSIPCO 2020 - Proceedings
PB - European Signal Processing Conference, EUSIPCO
T2 - 28th European Signal Processing Conference, EUSIPCO 2020
Y2 - 24 August 2020 through 28 August 2020
ER -