Analysis of Conventional, Near-Memory, and In-Memory DNN Accelerators
Author(s)
Abstract
Various DNN accelerators based on Conventional compute Hardware Accelerator (CHA), Near-Data-Processing (NDP) and Processing-in-Memory (PIM) paradigms have been proposed to meet the challenges of inferencing Deep Neural Networks (DNNs). To the best of our knowledge, this work aims to perform the first quantitative as well as qualitative comparison among the state-of-the-art accelerators from each digital DNN accelerator paradigm. Our study provides insights into selecting the best architecture for a given DNN workload. We have used workloads of the MLPerf Inference benchmark. We observe that for Fully Connected Layer (FCL) DNNs, PIM-based accelerator is 21� and 3� faster than CHA and NDP-based accelerator respectively. However, NDP is 9� and 2.5� more energy efficient than CHA and PIM for FCL. For Convolutional Neural Network (CNN) workloads, CHA is 10% and 5� faster than NDP and PIM-based accelerator respectively. Further, CHA is 1.5� and 6� more energy efficient than NDP and PIM-based accelerators respectively. � 2023 Elsevier B.V., All rights reserved.
Keywords
Convolutional neural networks
Data handling
Energy efficiency
Convolutional neural network
Data processors
Deep neural network accelerator
Energy efficient
Hardware accelerators
Memory paradigm
Near data processor
Neural-network processing
Processing-in-memory
State of the art
Deep neural networks
