https://scholars.lib.ntu.edu.tw/handle/123456789/607462
標題: | Intra- And Inter- Layer Transformation to Reduce Memory Traffic for CNN Computation | 作者: | Liao P.-W WEI-CHUNG HSU SHIH-WEI LIAO |
關鍵字: | Computer aided design;Scheduling algorithms;Static random access storage;Adjacent layers;Data reuse;Inter-layers;Intra-layer;Layer data;Layer transformation;Local memory;Off-chip;On chips;Processing elements;Dynamic random access storage | 公開日期: | 2021 | 來源出版物: | ACM International Conference Proceeding Series | 摘要: | Edge inference has gained much popularity in recent years. Many AI accelerators have been proposed and extensively studied. Such devices are often packed with a large number of PEs (Processing Elements), and lots of on-chip SRAM. The key to successful AI acceleration is to effectively use the data transferred from off-chip DRAM to the on-chip SRAM. Most prior studies optimize the use of on-chip SRAM for a single convolution layer, they tend to ignore the opportunity of inter-layer data reuse. We have proposed an algorithm to schedule two adjacent layers of CNN operations. Our goal is to reduce traffic between DRAM and local memory more than allocating the buffer to only a single layer. Our cross-layer scheduling effectively reduces the memory traffic. We hav also verified the validity of our memory traffic reduction model on the Gemmini simulator from UC Berkeley. ? 2021 ACM. |
URI: | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85115975157&doi=10.1145%2f3458744.3473353&partnerID=40&md5=b68ae2dfad5c8216bbb297f07a612b45 https://scholars.lib.ntu.edu.tw/handle/123456789/607462 |
DOI: | 10.1145/3458744.3473353 |
顯示於: | 資訊工程學系 |
在 IR 系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。