Please use this identifier to cite or link to this item: https://dspace.iiti.ac.in/handle/123456789/4875
Full metadata record
DC FieldValueLanguage
dc.contributor.authorGupta, Siddharthen_US
dc.contributor.authorAhuja, Kapilen_US
dc.contributor.authorTiwari, Arunaen_US
dc.contributor.authorKumar, Akashen_US
dc.date.accessioned2022-03-17T01:00:00Z-
dc.date.accessioned2022-03-17T15:35:50Z-
dc.date.available2022-03-17T01:00:00Z-
dc.date.available2022-03-17T15:35:50Z-
dc.date.issued2020-
dc.identifier.citationGupta, S., Ullah, S., Ahuja, K., Tiwari, A., & Kumar, A. (2020). ALigN: A highly accurate adaptive layerwise Log_2_Lead quantization of pre-trained neural networks. IEEE Access, 8, 118899-118911. doi:10.1109/ACCESS.2020.3005286en_US
dc.identifier.issn2169-3536-
dc.identifier.otherEID(2-s2.0-85088285380)-
dc.identifier.urihttps://doi.org/10.1109/ACCESS.2020.3005286-
dc.identifier.urihttps://dspace.iiti.ac.in/handle/123456789/4875-
dc.description.abstractDeep Neural Networks are one of the machine learning techniques which are increasingly used in a variety of applications. However, the significantly high memory and computation demands of deep neural networks often limit their deployment on embedded systems. Many recent works have considered this problem by proposing different types of data quantization schemes. However, most of these techniques either require post-quantization retraining of deep neural networks or bear a significant loss in output accuracy. In this paper, we propose a novel and scalable technique with two different modes for the quantization of the parameters of pre-trained neural networks. In the first mode, referred to as log_2_lead, we use a single template for the quantization of all parameters. In the second mode, denoted as ALigN, we analyze the trained parameters of each layer and adaptively adjust the quantization template to achieve even higher accuracy. Our technique significantly maintains the accuracy of the parameters and does not require retraining of the networks. Moreover, it supports quantization to an arbitrary bit-size. For example, compared to the single-precision floating-point numbers-based implementation, our proposed 8-bit quantization technique generates only ∼ 0.2% and ∼ 0.1% , loss in the Top-1 and Top-5 accuracies respectively for VGG-16 network using ImageNet dataset. We have observed similar minimal losses in the Top-1 and Top-5 accuracies for AlexNet and Resnet-18 using the proposed quantization scheme for the 8-bit range. Our proposed quantization technique also provides a higher mean intersection over union for semantic segmentation when compared with state-of-the-art quantization techniques. The proposed technique represents parameters in powers of 2, thereby eliminating the need for resource-computationally intensive multiplier units for the hardware accelerators of the neural networks. We also present a design for implementing the multiplication operation using bit-shifts and addition for the proposed quantization technique. © 2013 IEEE.en_US
dc.language.isoenen_US
dc.publisherInstitute of Electrical and Electronics Engineers Inc.en_US
dc.sourceIEEE Accessen_US
dc.subjectDeep neural networksen_US
dc.subjectDigital arithmeticen_US
dc.subjectEmbedded systemsen_US
dc.subjectLearning systemsen_US
dc.subjectSemanticsen_US
dc.subjectData quantizationsen_US
dc.subjectHardware acceleratorsen_US
dc.subjectMachine learning techniquesen_US
dc.subjectMultiplication operationsen_US
dc.subjectQuantization schemesen_US
dc.subjectSemantic segmentationen_US
dc.subjectState of the arten_US
dc.subjectTrained neural networksen_US
dc.subjectMultilayer neural networksen_US
dc.titleALigN: A Highly Accurate Adaptive Layerwise Log_2_Lead Quantization of Pre-Trained Neural Networksen_US
dc.typeJournal Articleen_US
dc.rights.licenseAll Open Access, Gold-
Appears in Collections:Department of Computer Science and Engineering

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetric Badge: