Abstract
The application of Artificial Intelligence (AI) to multiple sectors has grown impressively in the last decade, posing concerns about energy consumption and environmental footprint in this field. The approximate computing paradigm reports promising techniques for the design of Deep Neural Network (DNN) accelerators to reduce resource consumption in both low-power devices and large-scale inference. This work addresses the resource and power consumption challenge by proposing the implementation of configurable approximate arithmetic operators described in untimed C++ for High-Level Synthesis (HLS), evaluating the impact of the approximations on the model accuracy of Neural Networks (NN) used for classification with Zero-Shot Quantisation (ZSQ) and without fine-tuning. Our proposed operators are fully parametric in terms of the number of approximated bits and numerical precision by using C++ templated and achieve up to 39.04% resource savings while having 79% accuracy in a LeNet-5 trained for MNIST.
| Original language | English |
|---|---|
| Title of host publication | 2024 IEEE 42nd Central America and Panama Convention, CONCAPAN 2024 |
| Publisher | Institute of Electrical and Electronics Engineers Inc. |
| Edition | 2024 |
| ISBN (Electronic) | 9798350366723 |
| DOIs | |
| State | Published - 2024 |
| Event | 42nd IEEE Central America and Panama Convention, CONCAPAN 2024 - San Jose, Costa Rica Duration: 27 Nov 2024 → 29 Nov 2024 |
Conference
| Conference | 42nd IEEE Central America and Panama Convention, CONCAPAN 2024 |
|---|---|
| Country/Territory | Costa Rica |
| City | San Jose |
| Period | 27/11/24 → 29/11/24 |
UN SDGs
This output contributes to the following UN Sustainable Development Goals (SDGs)
-
SDG 7 Affordable and Clean Energy
-
SDG 12 Responsible Consumption and Production
Keywords
- Approximate computing
- field programmable gate arrays
- inference
- machine learning
- neural networks
Fingerprint
Dive into the research topics of 'Configurable High-Level Synthesis Approximate Arithmetic Units for Deep Learning Accelerators'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver