Row-wise Accelerator for Vision Transformer

Hong Yi Wang, Tian Sheuan Chang

研究成果: Conference contribution同行評審

3 引文 斯高帕斯(Scopus)

摘要

Following the success of the natural language processing, the transformer for vision applications has attracted significant attention in recent years due to its excellent performance. However, existing deep learning hardware accelerators for vision cannot execute this structure efficiently due to significant model architecture differences. As a result, this paper proposes the hardware accelerator for vision transformers with row-wise scheduling, which decomposes major operations in vision transformers as a single dot product primitive for a unified and efficient execution. Furthermore, by sharing weights in columns, we can reuse the data and reduce the usage of memory. The implementation with TSMC 40nm CMOS technology only requires 262K gate count and 149KB SRAM buffer for 403.2 GOPS throughput at 600MHz clock frequency.

原文English
主出版物標題Proceeding - IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022
發行者Institute of Electrical and Electronics Engineers Inc.
頁面399-402
頁數4
ISBN(電子)9781665409964
DOIs
出版狀態Published - 2022
事件4th IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022 - Incheon, Korea, Republic of
持續時間: 13 6月 202215 6月 2022

出版系列

名字Proceeding - IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022

Conference

Conference4th IEEE International Conference on Artificial Intelligence Circuits and Systems, AICAS 2022
國家/地區Korea, Republic of
城市Incheon
期間13/06/2215/06/22

指紋

深入研究「Row-wise Accelerator for Vision Transformer」主題。共同形成了獨特的指紋。

引用此