Please use this identifier to cite or link to this item: https://dspace.iiti.ac.in/handle/123456789/11296
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorAhuja, Kapil-
dc.contributor.advisorTiwari, Aruna-
dc.contributor.authorKumar, Saurav-
dc.date.accessioned2023-02-22T11:48:02Z-
dc.date.available2023-02-22T11:48:02Z-
dc.date.issued2023-02-08-
dc.identifier.urihttps://dspace.iiti.ac.in/handle/123456789/11296-
dc.description.abstractDeep neural networks (DNNs) are widely used and have demonstrated their usefulness in many applications, such as computer vision and pattern recognition. However, the training and inference of these networks can be time-consuming. Such a problem could be alleviated by using efficient optimizers. As one of the most commonly used opti mizers, Adagrad adaptively selects the learning rate using the gradient information. Although Adagrad works reasonably well, it does have drawbacks of overshoot and oscillations around the minima. This slows the convergence of Adagrad. To alleviate this overshoot problem and accelerate the convergence of DNN optimization, we combine the adaptiveness of Adagrad with the change in gradient information as in the PID optimizer. We term our new optimizer as a Control parameter-based gradient descent optimizer or CaaGrad. Using a Multi-Layered Perceptron (MLP) on the MNIST and Fashion MNIST datasets, we show the average accuracy in training, validation, and testing of CaaGrad is 9.96%, 8.67%, and 4.19% better than Adagrad, PID optimizer, and Adam (another state-of-the-art optimizer), respectively. To help overcome the oscillations in CaaGrad, we add a controlling factor to it. We term our new optimizer CaaGrad2. Using MLP (Multi-Layered Perceptron) on MNIST and Fashion MNIST datasets, the average accuracy (in training, validation, and testing) of CaaGrad is 9.11% better than that of Adam.en_US
dc.language.isoenen_US
dc.publisherDepartment of Computer Science and Engineering, IIT Indoreen_US
dc.relation.ispartofseriesMSR031;-
dc.subjectComputer Science and Engineeringen_US
dc.titleCaaGrad: an adaptive approach for optimization in deep learningen_US
dc.typeThesis_MS Researchen_US
Appears in Collections:Department of Computer Science and Engineering_ETD

Files in This Item:
File Description SizeFormat 
MSR031_Saurav_Kumar_2004101009.pdf2.61 MBAdobe PDFView/Open


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetric Badge: