Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning

Deep learning models are part of the family of artificial neural networks and, as such, they suffer catastrophic interference when learning sequentially. In addition, the greater number of these models have a rigid architecture which prevents the incremental learning of new classes. To overcome thes...

Full description

Bibliographic Details
Main Authors: Diego Mellado, Carolina Saavedra, Steren Chabert, Romina Torres, Rodrigo Salas
Format: Article
Language:English
Published: MDPI AG 2019-10-01
Series:Algorithms
Subjects:
Online Access:https://www.mdpi.com/1999-4893/12/10/206
id doaj-4c1b6cc1d3d0445f9ed98cd0778b41f3
record_format Article
spelling doaj-4c1b6cc1d3d0445f9ed98cd0778b41f32020-11-25T02:03:28ZengMDPI AGAlgorithms1999-48932019-10-01121020610.3390/a12100206a12100206Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class LearningDiego Mellado0Carolina Saavedra1Steren Chabert2Romina Torres3Rodrigo Salas4Escuela de Ingeniería C. Biomédica, Universidad de Valaraíso, Valparaíso 2362905, ChileEscuela de Ingeniería C. Biomédica, Universidad de Valaraíso, Valparaíso 2362905, ChileEscuela de Ingeniería C. Biomédica, Universidad de Valaraíso, Valparaíso 2362905, ChileEngineering Faculty, Universidad Andres Bello, Viña del Mar 2531015, ChileEscuela de Ingeniería C. Biomédica, Universidad de Valaraíso, Valparaíso 2362905, ChileDeep learning models are part of the family of artificial neural networks and, as such, they suffer catastrophic interference when learning sequentially. In addition, the greater number of these models have a rigid architecture which prevents the incremental learning of new classes. To overcome these drawbacks, we propose the Self-Improving Generative Artificial Neural Network (SIGANN), an end-to-end deep neural network system which can ease the catastrophic forgetting problem when learning new classes. In this method, we introduce a novel detection model that automatically detects samples of new classes, and an adversarial autoencoder is used to produce samples of previous classes. This system consists of three main modules: a classifier module implemented using a Deep Convolutional Neural Network, a generator module based on an adversarial autoencoder, and a novelty-detection module implemented using an OpenMax activation function. Using the EMNIST data set, the model was trained incrementally, starting with a small set of classes. The results of the simulation show that SIGANN can retain previous knowledge while incorporating gradual forgetfulness of each learning sequence at a rate of about 7% per training step. Moreover, SIGANN can detect new classes that are hidden in the data with a median accuracy of <inline-formula> <math display="inline"> <semantics> <mrow> <mn>43</mn> <mo>%</mo> </mrow> </semantics> </math> </inline-formula> and, therefore, proceed with incremental class learning.https://www.mdpi.com/1999-4893/12/10/206artificial neural networksdeep learninggenerative neural networksincremental learningnovelty detectioncatastrophic interference
collection DOAJ
language English
format Article
sources DOAJ
author Diego Mellado
Carolina Saavedra
Steren Chabert
Romina Torres
Rodrigo Salas
spellingShingle Diego Mellado
Carolina Saavedra
Steren Chabert
Romina Torres
Rodrigo Salas
Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
Algorithms
artificial neural networks
deep learning
generative neural networks
incremental learning
novelty detection
catastrophic interference
author_facet Diego Mellado
Carolina Saavedra
Steren Chabert
Romina Torres
Rodrigo Salas
author_sort Diego Mellado
title Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
title_short Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
title_full Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
title_fullStr Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
title_full_unstemmed Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
title_sort self-improving generative artificial neural network for pseudorehearsal incremental class learning
publisher MDPI AG
series Algorithms
issn 1999-4893
publishDate 2019-10-01
description Deep learning models are part of the family of artificial neural networks and, as such, they suffer catastrophic interference when learning sequentially. In addition, the greater number of these models have a rigid architecture which prevents the incremental learning of new classes. To overcome these drawbacks, we propose the Self-Improving Generative Artificial Neural Network (SIGANN), an end-to-end deep neural network system which can ease the catastrophic forgetting problem when learning new classes. In this method, we introduce a novel detection model that automatically detects samples of new classes, and an adversarial autoencoder is used to produce samples of previous classes. This system consists of three main modules: a classifier module implemented using a Deep Convolutional Neural Network, a generator module based on an adversarial autoencoder, and a novelty-detection module implemented using an OpenMax activation function. Using the EMNIST data set, the model was trained incrementally, starting with a small set of classes. The results of the simulation show that SIGANN can retain previous knowledge while incorporating gradual forgetfulness of each learning sequence at a rate of about 7% per training step. Moreover, SIGANN can detect new classes that are hidden in the data with a median accuracy of <inline-formula> <math display="inline"> <semantics> <mrow> <mn>43</mn> <mo>%</mo> </mrow> </semantics> </math> </inline-formula> and, therefore, proceed with incremental class learning.
topic artificial neural networks
deep learning
generative neural networks
incremental learning
novelty detection
catastrophic interference
url https://www.mdpi.com/1999-4893/12/10/206
work_keys_str_mv AT diegomellado selfimprovinggenerativeartificialneuralnetworkforpseudorehearsalincrementalclasslearning
AT carolinasaavedra selfimprovinggenerativeartificialneuralnetworkforpseudorehearsalincrementalclasslearning
AT sterenchabert selfimprovinggenerativeartificialneuralnetworkforpseudorehearsalincrementalclasslearning
AT rominatorres selfimprovinggenerativeartificialneuralnetworkforpseudorehearsalincrementalclasslearning
AT rodrigosalas selfimprovinggenerativeartificialneuralnetworkforpseudorehearsalincrementalclasslearning
_version_ 1724947986293719040