Please use this identifier to cite or link to this item: https://dspace.iiti.ac.in/handle/123456789/14550
Title: Unsupervised Continual Learning using Cross-Level Discrimination and Evidential Pseudo Out-of-Distribution Detection along with Gradient Projected Hard Attention
Authors: Malviya, Ankit
Maurya, Chandresh Kumar
Keywords: Accuracy;catastrophic forgetting;continual learning;Continuing education;Contrastive learning;Parameter estimation;parameter isolation;rehearsal free;Representation learning;Task analysis;Training;Uncertainty;Unsupervised learning;unsupervised learning
Issue Date: 2024
Publisher: Institute of Electrical and Electronics Engineers Inc.
Citation: Malviya, A., & Maurya, C. K. (2024). Unsupervised Continual Learning using Cross-Level Discrimination and Evidential Pseudo Out-of-Distribution Detection along with Gradient Projected Hard Attention. IEEE Access. Scopus. https://doi.org/10.1109/ACCESS.2024.3435555
Abstract: Catastrophic forgetting is a prominent challenge in machine learning. It denotes the phenomenon wherein models undergo a significant loss of previously acquired knowledge upon learning new information. Supervised Continual Learning (SCL) has emerged as a promising approach to mitigate this issue by enabling models to adapt to non-stationary data distributions while leveraging labeled data. However, practical limitations arise for SCL in real-world settings, where labeled data is scarce. Conversely, Unsupervised Continual Learning (UCL) aims to mitigate forgetting without the need for manual annotations. Still, many previous UCL methods rely on replay-based strategies, which may not be viable in contexts where storing training data is impractical. Additionally, replay methods may face challenges related to overfitting and representation drifts when the buffer size is limited. To overcome the limitations of replay strategies, we propose an approach based on parameter isolation, combined with gradient projection. Specifically, we use task-specific hard attention and gradient projection to constrain updates to parameters crucial for previous tasks. Furthermore, our approach offers advantages over architecture-based methodologies by avoiding the need for network expansion and allowing for sequential learning within a predefined network architecture. Contrastive learning-based unsupervised methods are effective in preserving representation continuity. However, the loss in contrastive learning may suffer from a decrease in the diversity of negative samples. To address this, we incorporate both direct instance grouping and discrimination at the cross-level for actual and pseudo groups to contrastively learn unsupervised representations.We employ evidential deep learning (EDL) on rotation augmentation-based pseudo groups to effectively identify OOD instances and learn distinct class representations. Through comprehensive experiments, we demonstrate that the proposed model achieves an overall-average accuracy of 77.82% for task incremental learning (TIL), and 67.39% for class incremental learning (CIL) across various benchmark datasets, encompassing both short and long sequences of tasks. Notably, our model achieves almost zero forgetting, outperforming state-of-the-art (SOTA) baseline accuracies of 73.33% and 60.8% for TIL and CIL, respectively. Additionally, our model demonstrates a significant improvement over the SOTA SCL baseline, achieving a 2.82% and 3.68% increase in average TIL and CIL accuracy, respectively, while substantially reducing forgetting from approximately 12.66% and 19.56% to nearly zero. Authors
URI: https://doi.org/10.1109/ACCESS.2024.3435555
https://dspace.iiti.ac.in/handle/123456789/14550
ISSN: 2169-3536
Type of Material: Journal Article
Appears in Collections:Department of Computer Science and Engineering

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetric Badge: