Meta Transferring for Deblurring

Po Sheng Liu, Fu Jen Tsai, Yan Tsung Peng, Chung Chi Tsai, Chia Wen Lin, Yen Yu Lin

研究成果: Paper同行評審

1 引文 斯高帕斯(Scopus)

摘要

Most previous deblurring methods were built with a generic model trained on blurred images and their sharp counterparts. However, these approaches might have sub-optimal deblurring results due to the domain gap between the training and test sets. This paper proposes a reblur-deblur meta-transferring scheme to realize test-time adaptation without using ground truth for dynamic scene deblurring. Since the ground truth is usually unavailable at inference time in a real-world scenario, we leverage the blurred input video to find and use relatively sharp patches as the pseudo ground truth. Furthermore, we propose a reblurring model to extract the homogenous blur from the blurred input and transfer it to the pseudo-sharps to obtain the corresponding pseudo-blurred patches for meta-learning and test-time adaptation with only a few gradient updates. Extensive experimental results show that our reblur-deblur meta-learning scheme can improve state-of-the-art deblurring models on the DVD, REDS, and RealBlur benchmark datasets.The source code is available at https://github.com/po-sheng/Meta_Transferring_for_Deblurring.

原文English
出版狀態Published - 2022
事件33rd British Machine Vision Conference Proceedings, BMVC 2022 - London, 英國
持續時間: 21 11月 202224 11月 2022

Conference

Conference33rd British Machine Vision Conference Proceedings, BMVC 2022
國家/地區英國
城市London
期間21/11/2224/11/22

指紋

深入研究「Meta Transferring for Deblurring」主題。共同形成了獨特的指紋。

引用此