Self-Net: Lifelong Learning via Continual Self-Modeling

Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for eac...

Full description

Bibliographic Details
Main Authors: Jaya Krishna Mandivarapu, Blake Camp, Rolando Estrada
Format: Article
Language:English
Published: Frontiers Media S.A. 2020-04-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/article/10.3389/frai.2020.00019/full
id doaj-423bf60dc638479ab452889716e3b058
record_format Article
spelling doaj-423bf60dc638479ab452889716e3b0582020-11-25T02:05:16ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122020-04-01310.3389/frai.2020.00019502426Self-Net: Lifelong Learning via Continual Self-ModelingJaya Krishna MandivarapuBlake CampRolando EstradaLearning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning.https://www.frontiersin.org/article/10.3389/frai.2020.00019/fulldeep learningcontinual learningautoencodersmanifold learningcatastrophic forgetting
collection DOAJ
language English
format Article
sources DOAJ
author Jaya Krishna Mandivarapu
Blake Camp
Rolando Estrada
spellingShingle Jaya Krishna Mandivarapu
Blake Camp
Rolando Estrada
Self-Net: Lifelong Learning via Continual Self-Modeling
Frontiers in Artificial Intelligence
deep learning
continual learning
autoencoders
manifold learning
catastrophic forgetting
author_facet Jaya Krishna Mandivarapu
Blake Camp
Rolando Estrada
author_sort Jaya Krishna Mandivarapu
title Self-Net: Lifelong Learning via Continual Self-Modeling
title_short Self-Net: Lifelong Learning via Continual Self-Modeling
title_full Self-Net: Lifelong Learning via Continual Self-Modeling
title_fullStr Self-Net: Lifelong Learning via Continual Self-Modeling
title_full_unstemmed Self-Net: Lifelong Learning via Continual Self-Modeling
title_sort self-net: lifelong learning via continual self-modeling
publisher Frontiers Media S.A.
series Frontiers in Artificial Intelligence
issn 2624-8212
publishDate 2020-04-01
description Learning a set of tasks over time, also known as continual learning (CL), is one of the most challenging problems in artificial intelligence. While recent approaches achieve some degree of CL in deep neural networks, they either (1) store a new network (or an equivalent number of parameters) for each new task, (2) store training data from previous tasks, or (3) restrict the network's ability to learn new tasks. To address these issues, we propose a novel framework, Self-Net, that uses an autoencoder to learn a set of low-dimensional representations of the weights learned for different tasks. We demonstrate that these low-dimensional vectors can then be used to generate high-fidelity recollections of the original weights. Self-Net can incorporate new tasks over time with little retraining, minimal loss in performance for older tasks, and without storing prior training data. We show that our technique achieves over 10X storage compression in a continual fashion, and that it outperforms state-of-the-art approaches on numerous datasets, including continual versions of MNIST, CIFAR10, CIFAR100, Atari, and task-incremental CORe50. To the best of our knowledge, we are the first to use autoencoders to sequentially encode sets of network weights to enable continual learning.
topic deep learning
continual learning
autoencoders
manifold learning
catastrophic forgetting
url https://www.frontiersin.org/article/10.3389/frai.2020.00019/full
work_keys_str_mv AT jayakrishnamandivarapu selfnetlifelonglearningviacontinualselfmodeling
AT blakecamp selfnetlifelonglearningviacontinualselfmodeling
AT rolandoestrada selfnetlifelonglearningviacontinualselfmodeling
_version_ 1724939121207541760