Hardware-Software Codesign of DNN Accelerators Using Approximate Posit Multipliers
Source
Proceedings of the Asia and South Pacific Design Automation Conference ASP DAC
Date Issued
2023-01-16
Author(s)
Glint, Tom
Prasad, Kailash
Dagli, Jinay
Gandhi, Krishil
Gupta, Aryan
Patel, Vrajesh
Shah, Neel
Abstract
Emerging data intensive AI/ML workloads encounter memory and power wall when run on general-purpose compute cores. This has led to the development of a myriad of techniques to deal with such workloads, among which DNN accelerator architectures have found a prominent place. In this work, we propose a hardware-software co-design approach to achieve system-level benefits. We propose a quantized data-aware POSIT number representation that leads to a highly optimized DNN accelerator. We demonstrate this work on SOTA SIMBA architecture, extendable to any other accelerator. Our proposal reduces the buffer/storage requirements within the architecture and reduces the data transfer cost between the main memory and the DNN accelerator. We have investigated the impact of using integer, IEEE floating point, and posit multipliers for LeNet, ResNet and VGG NNs trained and tested on MNIST, CIFAR10 and ImageNet datasets, respectively. Our system-level analysis shows that the proposed approximate-fixed-posit multiplier when implemented on SIMBA architecture, achieves on average ∼2.2× speed up, consumes ∼3.1× less energy and requires ∼3.2× less area, respectively, against the baseline SOTA architecture, without loss of accuracy (∼±1%).
Subjects
co-design | DNN accelerators | neural networks
