High Utilization Energy-Aware Real-Time Inference Deep Convolutional Neural Network Accelerator

التفاصيل البيبلوغرافية
العنوان: High Utilization Energy-Aware Real-Time Inference Deep Convolutional Neural Network Accelerator
المؤلفون: Ching-Te Chiu, Kuan-Ting Lin, Shan-Chien Hsiao, Jheng-Yi Chang
المصدر: ISCAS
بيانات النشر: IEEE, 2021.
سنة النشر: 2021
مصطلحات موضوعية: Data access, Speedup, Computer engineering, Edge device, Data exchange, Computer science, Reuse, Chip, Convolutional neural network, Efficient energy use
الوصف: Deep convolution Neural Network (DCNN) has been widely used in computer vision tasks. However, for edge device, even then inference has too large computational complexity and data access amount. Due to the mentioned shortcomings, the inference latency of state-of-the-art models are still impractical for real-world applications. In this paper, we proposed a high utilization energy-aware real-time inference deep convolutional neural network accelerator, which outperforms the current accelerators. First, we use 1x1 size convolution kernels as the smallest unit of the computing unit. And we design suitable computing unit for different models based on the requirement of each model. Second, we use Reuse Feature SRAM to store the output of current layer in the chip and use as the input of the next layer. Moreover, we import Output Reuse Strategy and Ring Stream Data flow not only to expand the reuse rate of data in the chip but to reduce the amount of data exchange between chips and DRAM. Finally, we present On-fly Pooling Module to let the calculation of the Pooling layer to be completed directly in the chip. With the aid of the proposed method in this paper, the implemented CNN acceleration chip has extreme high hardware utilization rate. We reduce a generous amount of data transfer on the specific module, ECNN [1]. Compared to the methods without reuse strategy, we can reduce 533 times of data access amount. At the same time, we have enough computing power to perform real-time execution of the existing image classification model, VGG16 [2] and MobileNet [3]. Compared with the design in [4], we can speed up 7.52 times and have 1.92x energy efficiency.
DOI: 10.1109/iscas51556.2021.9401526
URL الوصول: https://explore.openaire.eu/search/publication?articleId=doi_________::2702aa7a6cf41b3ae8b1c4aac72c9046
https://doi.org/10.1109/iscas51556.2021.9401526
Rights: CLOSED
رقم الانضمام: edsair.doi...........2702aa7a6cf41b3ae8b1c4aac72c9046
قاعدة البيانات: OpenAIRE
الوصف
DOI:10.1109/iscas51556.2021.9401526