Please use this identifier to cite or link to this item:
https://dspace.iiti.ac.in/handle/123456789/2694
Title: | Approximated deep neural networks : A computationally efficient inference |
Authors: | Gupta, Siddharth |
Supervisors: | Ahuja, Kapil Tiwari, Aruna |
Keywords: | Computer Science and Engineering |
Issue Date: | 4-Dec-2020 |
Publisher: | Department of Computer Science and Engineering, IIT Indore |
Series/Report no.: | MSR001 |
Abstract: | For most of the pattern recognition and machine learning tasks, deep learning is performing well in recent years. Solving these machine learning tasks require large size deep neural networks (DNNs). Many state-of-the-art DNNs consists of millions of training parameters and arithmetic operations. Increase in size of deep neural network models results in high computational and memory space needs. Therefore, realization of these models on low power devices is possible with some approximation while retaining the accuracy of network. Many recent works have proposed different data quantization techniques. Most of these works require retraining of DNNs after quantization for reducing the accuracy loss due to quantization. Other works which do not retrain the network after quantization, suffer loss in accuracy. This thesis presents a scalable technique for representing trained parameters of deep neural networks in such a way that these trained parameters can be used for computational and memory-efficient inference phase of deep neural networks. This technique consists of two variations i.e. log 2 lead, and ALigN . log 2 lead provides a single template for parameter representation and ALigN adaptively adjusts the template according to different layers of DNN for producing even better results. We have taken three different DNNs, AlexNet, VGG-16, and Resnet-18, and we quantize them using 8-bit version of our schemes and found minimal loss in accuracy compared to full precision network. For evaluating the efficiency of our technique, we have applied our proposed techniques for image classification and segmentation tasks. We have also presented a multiplier design for efficient multiplication of values represented in our proposed templates. Keywords: Machine learning, deep neural networks, quantization, multipliers, classification, segmentation |
URI: | https://dspace.iiti.ac.in/handle/123456789/2694 |
Type of Material: | Thesis_MS Research |
Appears in Collections: | Department of Computer Science and Engineering_ETD |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
MSR001_Siddharth_Gupta_1804101006.pdf | 7.08 MB | Adobe PDF | ![]() View/Open |
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.
Altmetric Badge: