TY - GEN
T1 - Using Conditional Video Compressors for Image Restoration
AU - Chen, Yi Hsin
AU - Ho, Yen Kuan
AU - Lin, Ting Han
AU - Peng, Wen Hsiao
AU - Huang, Ching Chun
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - To address the ill-posed nature of image restoration tasks, recent research efforts have been focused on integrating conditional generative models, such as conditional variational autoencoders (CVAE). However, how to condition the autoencoder to maximize the conditional evidence lower bound remains an open issue, particularly for the restoration tasks. Inspired by the rapid advancements in CVAE-based video compression, we make the first attempt to adapt a conditional video compressor for image restoration. In doing so, we have the low-quality image to be enhanced, which plays the same role as the reference frame for conditional video coding. Our scheme applies scalar quantization in training the autoencoder, circumventing the difficulties of training a large-size codebook as with prior works that adopt vector-quantized VAE (VQ-VAE). Moreover, it trains end-to-end a fully conditioned autoencoder, including a conditional encoder, a conditional decoder, and a conditional prior network, to maximize the conditional evidence lower bound. Extensive experiments confirm the superiority of our scheme on denoising and deblurring tasks.
AB - To address the ill-posed nature of image restoration tasks, recent research efforts have been focused on integrating conditional generative models, such as conditional variational autoencoders (CVAE). However, how to condition the autoencoder to maximize the conditional evidence lower bound remains an open issue, particularly for the restoration tasks. Inspired by the rapid advancements in CVAE-based video compression, we make the first attempt to adapt a conditional video compressor for image restoration. In doing so, we have the low-quality image to be enhanced, which plays the same role as the reference frame for conditional video coding. Our scheme applies scalar quantization in training the autoencoder, circumventing the difficulties of training a large-size codebook as with prior works that adopt vector-quantized VAE (VQ-VAE). Moreover, it trains end-to-end a fully conditioned autoencoder, including a conditional encoder, a conditional decoder, and a conditional prior network, to maximize the conditional evidence lower bound. Extensive experiments confirm the superiority of our scheme on denoising and deblurring tasks.
KW - image deblurring
KW - image denoising
KW - image restoration
KW - learning-based video compression
UR - http://www.scopus.com/inward/record.url?scp=85215699001&partnerID=8YFLogxK
U2 - 10.1109/WOCC61718.2024.10786066
DO - 10.1109/WOCC61718.2024.10786066
M3 - Conference contribution
AN - SCOPUS:85215699001
T3 - 2024 33rd Wireless and Optical Communications Conference, WOCC 2024
SP - 154
EP - 158
BT - 2024 33rd Wireless and Optical Communications Conference, WOCC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 33rd Wireless and Optical Communications Conference, WOCC 2024
Y2 - 25 October 2024 through 26 October 2024
ER -