Imperceptible adversarial attack via spectral sensitivity of human visual system

Chen Kuo Chiang*, Ying Dar Lin, Ren Hung Hwang, Po Ching Lin, Shih Ya Chang, Hao Ting Li

*此作品的通信作者

研究成果: Article同行評審

1 引文 斯高帕斯(Scopus)

摘要

Adversarial attacks reveals that deep neural networks are vulnerable to adversarial examples. Intuitively, adversarial examples with more perturbations result in a strong attack, leading to a lower recognition accuracy. However, increasing perturbations also causes visually noticeable changes in the images. In order to address the problem on how to improve the attack strength while maintaining the visual perception quality, an imperceptible adversarial attack via spectral sensitivity of the human visual system is proposed. Based on the analysis of human visual system, the proposed method allows more perturbations as attack information and re-distributes perturbations into pixels where the changes are imperceptible to human eyes. Therefore, it presents better Accuracy under Attack(AuA) than existing attack methods whereas the image quality can be maintained to the similar level as other methods. Experimental results demonstrate that our method improves the attack strength of existing adversarial attack methods by adding 3% to 23% while mostly maintaining the visual quality of SSIM lower than 0.05.

原文English
頁(從 - 到)59291-59315
頁數25
期刊Multimedia Tools and Applications
83
發行號20
DOIs
出版狀態Published - 6月 2024

指紋

深入研究「Imperceptible adversarial attack via spectral sensitivity of human visual system」主題。共同形成了獨特的指紋。

引用此