Repository logo
  • English
  • العربية
  • বাংলা
  • Català
  • Čeština
  • Deutsch
  • Ελληνικά
  • Español
  • Suomi
  • Français
  • Gàidhlig
  • हिंदी
  • Magyar
  • Italiano
  • Қазақ
  • Latviešu
  • Nederlands
  • Polski
  • Português
  • Português do Brasil
  • Srpski (lat)
  • Српски
  • Svenska
  • Türkçe
  • Yкраї́нська
  • Tiếng Việt
Log In
New user? Click here to register.Have you forgotten your password?
  1. Home
  2. Scholalry Output
  3. Publications
  4. Hardware-Software Codesign of DNN Accelerators Using Approximate Posit Multipliers
 
  • Details

Hardware-Software Codesign of DNN Accelerators Using Approximate Posit Multipliers

Source
Proceedings of the Asia and South Pacific Design Automation Conference ASP DAC
Date Issued
2023-01-16
Author(s)
Glint, Tom
Prasad, Kailash
Dagli, Jinay
Gandhi, Krishil
Gupta, Aryan
Patel, Vrajesh
Shah, Neel
Mekie, Joycee  
DOI
10.1145/3566097.3567866
Abstract
Emerging data intensive AI/ML workloads encounter memory and power wall when run on general-purpose compute cores. This has led to the development of a myriad of techniques to deal with such workloads, among which DNN accelerator architectures have found a prominent place. In this work, we propose a hardware-software co-design approach to achieve system-level benefits. We propose a quantized data-aware POSIT number representation that leads to a highly optimized DNN accelerator. We demonstrate this work on SOTA SIMBA architecture, extendable to any other accelerator. Our proposal reduces the buffer/storage requirements within the architecture and reduces the data transfer cost between the main memory and the DNN accelerator. We have investigated the impact of using integer, IEEE floating point, and posit multipliers for LeNet, ResNet and VGG NNs trained and tested on MNIST, CIFAR10 and ImageNet datasets, respectively. Our system-level analysis shows that the proposed approximate-fixed-posit multiplier when implemented on SIMBA architecture, achieves on average ∼2.2× speed up, consumes ∼3.1× less energy and requires ∼3.2× less area, respectively, against the baseline SOTA architecture, without loss of accuracy (∼±1%).
Unpaywall
URI
https://d8.irins.org/handle/IITG2025/26920
Subjects
co-design | DNN accelerators | neural networks
IITGN Knowledge Repository Developed and Managed by Library

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Privacy policy
  • End User Agreement
  • Send Feedback
Repository logo COAR Notify