Abstract
Low latency inference has many applications in edge machine learning. In this paper, we present a run-time configurable convolutional neural network (CNN) inference ASIC design for low-latency edge machine learning. By implementing a 5-stage pipelined CNN inference model in a 3D ASIC technology, we demonstrate that the model distributed on two dies utilizing face-to-face (F2F) 3D integration achieves superior performance. Our experimental results show that the design based on 3D integration achieves 43% better energy-delay product when compared to the traditional 2D technology.
Original language | English |
---|---|
Title of host publication | ISCAS 2023 - 56th IEEE International Symposium on Circuits and Systems, Proceedings |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
ISBN (Electronic) | 9781665451093 |
DOIs | |
State | Published - 2023 |
Event | 56th IEEE International Symposium on Circuits and Systems, ISCAS 2023 - Monterey, United States Duration: May 21 2023 → May 25 2023 |
Publication series
Name | Proceedings - IEEE International Symposium on Circuits and Systems |
---|---|
Volume | 2023-May |
ISSN (Print) | 0271-4310 |
Conference
Conference | 56th IEEE International Symposium on Circuits and Systems, ISCAS 2023 |
---|---|
Country/Territory | United States |
City | Monterey |
Period | 05/21/23 → 05/25/23 |
Funding
V. ACKNOWLEDGEMENT This manuscript has been authored by UT-Battelle, LLC, under contract DE-AC05-00OR22725 with the US Department of Energy (DOE). The US government retains and the publisher, by accepting the article for publication, acknowledges that the US government retains a nonexclusive, paid-up, irrevocable, worldwide license to publish or reproduce the published form of this manuscript, or allow others to do so, for US government purposes. DOE will provide public access to these results of federally sponsored research in accordance with the DOE Public Access Plan (http://energy.gov/downloads/doe-public-access-plan).