Criteria for Learning without Forgetting in Artificial Neural Networks

Rupesh Karn, Prabhakar Kudva, Ibrahim Elfadel

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

7 Scopus citations

Abstract

Task progressive learning without catastrophic forgetting using artificial neural networks (ANNs) has demonstrated viability and promise. Due to the large number of ANN hyper-parameters, a model already trained over a group of tasks can further learn a new task without forgetting the previous tasks. Several algorithms have been proposed for progressive learning, including synaptic weight consolidation, ensemble, rehearsal, and sparse coding. One major problem with such methods is that they fail to detect the congestion in the ANN shared parameter space to indicate the saturation of the existing network and its inability to add new tasks using progressive learning. The detection of such saturation is especially needed to avoid the catastrophic forgetting of old trained task and the concurrent loss in their generalization quality. In this paper, we address such problem and propose a methodology for ANN congestion detection. The methodology is based on computing the Hessian of the ANN loss function at the optimal weights for a group of previously learned tasks. Since the Hessian calculation is compute-intensive, we provide a set of approximation heuristics that are computationally efficient. The algorithms are implemented and analyzed in the context of two cloud network security datasets, namely, UNSW-NB15 and AWID, as well as the MNIST image recognition dataset. Results show that the proposed congestion metrics give an accurate assessment of the ANN progressive learning capacity for these various datasets. Furthermore, the results show that models that have more features exhibit higher congestion thresholds and are therefore more amenable to progressive learning.

Original languageBritish English
Title of host publicationProceedings - 2019 IEEE International Conference on Cognitive Computing, ICCC 2019 - Part of the 2019 IEEE World Congress on Services
EditorsElisa Bertino, Carl K. Chang, Peter Chen, Ernesto Damiani, Michael Goul, Katsunori Oyama
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages90-97
Number of pages8
ISBN (Electronic)9781728127118
DOIs
StatePublished - Jul 2019
Event4th IEEE International Conference on Cognitive Computing, ICCC 2019 - Milan, Italy
Duration: 8 Jul 201913 Jul 2019

Publication series

NameProceedings - 2019 IEEE International Conference on Cognitive Computing, ICCC 2019 - Part of the 2019 IEEE World Congress on Services

Conference

Conference4th IEEE International Conference on Cognitive Computing, ICCC 2019
Country/TerritoryItaly
CityMilan
Period8/07/1913/07/19

Keywords

  • Catastrophic forgetting
  • Hessian
  • Neural Network
  • Progressive Learning

Fingerprint

Dive into the research topics of 'Criteria for Learning without Forgetting in Artificial Neural Networks'. Together they form a unique fingerprint.

Cite this