TY - GEN
T1 - BACF-Net
T2 - 2025 IEEE International Symposium on Circuits and Systems, ISCAS 2025
AU - Chang, Chen Lin
AU - Hsiao, Hsu Feng
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - In the era of big data, efficient image compression is essential for managing the rapid growth of visual content. While traditional codecs have advanced significantly, their dependence on handcrafted techniques limits their effectiveness. The emergence of deep learning has transformed image compression by enabling end-to-end optimization. However, existing learned methods typically utilize convolutional neural networks for local modeling or transformers for capturing long-range dependencies, indicating potential areas for enhancement. We introduce a novel hybrid hyperprior-based architecture that combines the advantages of residual CNNs and transformers to improve rate-distortion performance in learned image compression. Additionally, our proposed Bifurcated Attention-Convolution Fusion (BACF) block employs a parallel configuration of an enhanced residual CNN with split attention alongside mixed transformer variants for multi-axis attention and shifted window-based attention. This design allows the network to effectively process and integrate both local details and high-level semantic information. Extensive experiments on the Kodak, CLIC, and Tecnick datasets show that our proposed method achieves competitive rate-distortion performance.
AB - In the era of big data, efficient image compression is essential for managing the rapid growth of visual content. While traditional codecs have advanced significantly, their dependence on handcrafted techniques limits their effectiveness. The emergence of deep learning has transformed image compression by enabling end-to-end optimization. However, existing learned methods typically utilize convolutional neural networks for local modeling or transformers for capturing long-range dependencies, indicating potential areas for enhancement. We introduce a novel hybrid hyperprior-based architecture that combines the advantages of residual CNNs and transformers to improve rate-distortion performance in learned image compression. Additionally, our proposed Bifurcated Attention-Convolution Fusion (BACF) block employs a parallel configuration of an enhanced residual CNN with split attention alongside mixed transformer variants for multi-axis attention and shifted window-based attention. This design allows the network to effectively process and integrate both local details and high-level semantic information. Extensive experiments on the Kodak, CLIC, and Tecnick datasets show that our proposed method achieves competitive rate-distortion performance.
UR - https://www.scopus.com/pages/publications/105010644042
U2 - 10.1109/ISCAS56072.2025.11043965
DO - 10.1109/ISCAS56072.2025.11043965
M3 - Conference contribution
AN - SCOPUS:105010644042
T3 - Proceedings - IEEE International Symposium on Circuits and Systems
BT - ISCAS 2025 - IEEE International Symposium on Circuits and Systems, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
Y2 - 25 May 2025 through 28 May 2025
ER -