Please use this identifier to cite or link to this item: https://dspace.iiti.ac.in/handle/123456789/13610
Full metadata record
DC FieldValueLanguage
dc.contributor.authorAshar, Nehaen_US
dc.contributor.authorRaut, Gopalen_US
dc.contributor.authorTrivedi, Vasundharaen_US
dc.contributor.authorVishvakarma, Santosh Kumaren_US
dc.date.accessioned2024-04-26T12:43:28Z-
dc.date.available2024-04-26T12:43:28Z-
dc.date.issued2024-
dc.identifier.citationAshar, N., Raut, G., Trivedi, V., Vishvakarma, S. K., & Kumar, A. (2024). QuantMAC: Enhancing Hardware Performance in DNNs with Quantize Enabled Multiply-Accumulate Unit. IEEE Access. Scopus. https://doi.org/10.1109/ACCESS.2024.3379906en_US
dc.identifier.issn2169-3536-
dc.identifier.otherEID(2-s2.0-85188432106)-
dc.identifier.urihttps://doi.org/10.1109/ACCESS.2024.3379906-
dc.identifier.urihttps://dspace.iiti.ac.in/handle/123456789/13610-
dc.description.abstractIn response to the escalating demand for hardware-efficient Deep Neural Network (DNN) architectures, we present a novel quantize-enabled multiply-accumulate (MAC) unit. Our methodology employs a right shift-and-add computation for MAC operation, enabling runtime truncation without additional hardware. This architecture optimally utilizes hardware resources, enhancing throughput performance while reducing computational complexity through bit-truncation techniques. Our key methodology involves designing a hardware-efficient MAC computational algorithm that supports both iterative and pipeline implementations, catering to diverse hardware efficiency or enhanced throughput requirements in accelerators. Additionally, we introduce a processing element (PE) with a pre-loading bias scheme, reducing one clock delay and eliminating the need for conventional extra resources in PE implementation. The PE facilitates quantization-based MAC calculations through an efficient bit-truncation method, removing the necessity for extra hardware logic. This versatile PE accommodates variable bit-precision with a dynamic fraction part within the sfxpt< N,f > representation, meeting specific model or layer demands. Through software emulation, our proposed approach demonstrates minimal accuracy loss, revealing under 1.6% loss for LeNet-5 using MNIST and around 4% for ResNet-18 and VGG-16 with CIFAR-10 in the sfxpt< 8 ,5 > format compared to conventional float32-based implementations. Hardware performance parameters on the Xilinx-Virtex-7 board unveil a 37% reduction in area utilization and a 45% reduction in power consumption compared to the best state-of-the-art MAC architecture. Extending the proposed MAC to a LeNet DNN model results in a 42% reduction in resource requirements and a significant 27% reduction in delay. This architecture provides notable advantages for resource-efficient, high-throughput edge-AI applications. � 2013 IEEE.en_US
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.sourceIEEE Accessen_US
dc.subjectApproximate computeen_US
dc.subjectbit-truncationen_US
dc.subjectCORDICen_US
dc.subjectdeep neural networken_US
dc.subjecthardware acceleratoren_US
dc.subjectquantize processing elementen_US
dc.titleQuantMAC: Enhancing Hardware Performance in DNNs with Quantize Enabled Multiply-Accumulate Uniten_US
dc.typeJournal Articleen_US
dc.rights.licenseAll Open Access, Gold-
Appears in Collections:Department of Electrical Engineering

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetric Badge: