https://scholars.lib.ntu.edu.tw/handle/123456789/607462
DC 欄位 | 值 | 語言 |
---|---|---|
dc.contributor.author | Liao P.-W | en_US |
dc.contributor.author | WEI-CHUNG HSU | en_US |
dc.contributor.author | SHIH-WEI LIAO | en_US |
dc.creator | Liao P.-W;Hsu W.-C;Liao S.-W. | - |
dc.date.accessioned | 2022-04-25T06:43:59Z | - |
dc.date.available | 2022-04-25T06:43:59Z | - |
dc.date.issued | 2021 | - |
dc.identifier.uri | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85115975157&doi=10.1145%2f3458744.3473353&partnerID=40&md5=b68ae2dfad5c8216bbb297f07a612b45 | - |
dc.identifier.uri | https://scholars.lib.ntu.edu.tw/handle/123456789/607462 | - |
dc.description.abstract | Edge inference has gained much popularity in recent years. Many AI accelerators have been proposed and extensively studied. Such devices are often packed with a large number of PEs (Processing Elements), and lots of on-chip SRAM. The key to successful AI acceleration is to effectively use the data transferred from off-chip DRAM to the on-chip SRAM. Most prior studies optimize the use of on-chip SRAM for a single convolution layer, they tend to ignore the opportunity of inter-layer data reuse. We have proposed an algorithm to schedule two adjacent layers of CNN operations. Our goal is to reduce traffic between DRAM and local memory more than allocating the buffer to only a single layer. Our cross-layer scheduling effectively reduces the memory traffic. We hav also verified the validity of our memory traffic reduction model on the Gemmini simulator from UC Berkeley. ? 2021 ACM. | en_US |
dc.relation.ispartof | ACM International Conference Proceeding Series | en_US |
dc.subject | Computer aided design | en_US |
dc.subject | Scheduling algorithms | en_US |
dc.subject | Static random access storage | en_US |
dc.subject | Adjacent layers | en_US |
dc.subject | Data reuse | en_US |
dc.subject | Inter-layers | en_US |
dc.subject | Intra-layer | en_US |
dc.subject | Layer data | en_US |
dc.subject | Layer transformation | en_US |
dc.subject | Local memory | en_US |
dc.subject | Off-chip | en_US |
dc.subject | On chips | en_US |
dc.subject | Processing elements | en_US |
dc.subject | Dynamic random access storage | en_US |
dc.title | Intra- And Inter- Layer Transformation to Reduce Memory Traffic for CNN Computation | en_US |
dc.type | conference paper | en_US |
dc.identifier.doi | 10.1145/3458744.3473353 | - |
dc.identifier.scopus | 2-s2.0-85115975157 | - |
item.openairecristype | http://purl.org/coar/resource_type/c_5794 | - |
item.openairetype | conference paper | - |
item.grantfulltext | none | - |
item.cerifentitytype | Publications | - |
item.fulltext | no fulltext | - |
crisitem.author.dept | Computer Science and Information Engineering | - |
crisitem.author.dept | MediaTek-NTU Research Center | - |
crisitem.author.dept | Networking and Multimedia | - |
crisitem.author.dept | Computer Science and Information Engineering | - |
crisitem.author.dept | MediaTek-NTU Research Center | - |
crisitem.author.orcid | 0000-0002-0833-7981 | - |
crisitem.author.orcid | 0000-0001-5294-5274 | - |
crisitem.author.parentorg | College of Electrical Engineering and Computer Science | - |
crisitem.author.parentorg | Others: University-Level Research Centers | - |
crisitem.author.parentorg | College of Electrical Engineering and Computer Science | - |
crisitem.author.parentorg | College of Electrical Engineering and Computer Science | - |
crisitem.author.parentorg | Others: University-Level Research Centers | - |
顯示於: | 資訊工程學系 |
在 IR 系統中的文件,除了特別指名其著作權條款之外,均受到著作權保護,並且保留所有的權利。