Continual Learning for Deep Dense Prediction

Transferring a deep learning model from old tasks to a new one is known to suffer from the catastrophic forgetting effects. Such forgetting mechanism is problematic as it does not allow us to accumulate knowledge sequentially and requires retaining and retraining on all the training data. Existing t...

Full description

Bibliographic Details
Main Author: Lokegaonkar, Sanket Avinash
Other Authors: Computer Science
Format: Others
Published: Virginia Tech 2018
Subjects:
Online Access:http://hdl.handle.net/10919/83513
id ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-83513
record_format oai_dc
spelling ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-835132020-11-12T05:42:55Z Continual Learning for Deep Dense Prediction Lokegaonkar, Sanket Avinash Computer Science Ramakrishnan, Naren Huang, Jia-Bin Huang, Bert Computer Vision Continual Learning Image Segmentation Dense Prediction Transferring a deep learning model from old tasks to a new one is known to suffer from the catastrophic forgetting effects. Such forgetting mechanism is problematic as it does not allow us to accumulate knowledge sequentially and requires retaining and retraining on all the training data. Existing techniques for mitigating the abrupt performance degradation on previously trained tasks are mainly studied in the context of image classification. In this work, we present a simple method to alleviate catastrophic forgetting for pixel-wise dense labeling problems. We build upon the regularization technique using knowledge distillation to minimize the discrepancy between the posterior distribution of pixel class labels for old tasks predicted from 1) the original and 2) the updated networks. This technique, however, might fail in circumstances where the source and target distribution differ significantly. To handle the above scenario, we further propose an improvement to the distillation based approach by adding adaptive l2-regularization depending upon the per-parameter importance to the older tasks. We train our model on FCN8s, but our training can be generalized to stronger models like DeepLab, PSPNet, etc. Through extensive evaluation and comparisons, we show that our technique can incrementally train dense prediction models for novel object classes, different visual domains, and different visual tasks. Master of Science 2018-06-12T08:00:31Z 2018-06-12T08:00:31Z 2018-06-11 Thesis vt_gsexam:16094 http://hdl.handle.net/10919/83513 In Copyright http://rightsstatements.org/vocab/InC/1.0/ ETD application/pdf Virginia Tech
collection NDLTD
format Others
sources NDLTD
topic Computer Vision
Continual Learning
Image Segmentation
Dense Prediction
spellingShingle Computer Vision
Continual Learning
Image Segmentation
Dense Prediction
Lokegaonkar, Sanket Avinash
Continual Learning for Deep Dense Prediction
description Transferring a deep learning model from old tasks to a new one is known to suffer from the catastrophic forgetting effects. Such forgetting mechanism is problematic as it does not allow us to accumulate knowledge sequentially and requires retaining and retraining on all the training data. Existing techniques for mitigating the abrupt performance degradation on previously trained tasks are mainly studied in the context of image classification. In this work, we present a simple method to alleviate catastrophic forgetting for pixel-wise dense labeling problems. We build upon the regularization technique using knowledge distillation to minimize the discrepancy between the posterior distribution of pixel class labels for old tasks predicted from 1) the original and 2) the updated networks. This technique, however, might fail in circumstances where the source and target distribution differ significantly. To handle the above scenario, we further propose an improvement to the distillation based approach by adding adaptive l2-regularization depending upon the per-parameter importance to the older tasks. We train our model on FCN8s, but our training can be generalized to stronger models like DeepLab, PSPNet, etc. Through extensive evaluation and comparisons, we show that our technique can incrementally train dense prediction models for novel object classes, different visual domains, and different visual tasks. === Master of Science
author2 Computer Science
author_facet Computer Science
Lokegaonkar, Sanket Avinash
author Lokegaonkar, Sanket Avinash
author_sort Lokegaonkar, Sanket Avinash
title Continual Learning for Deep Dense Prediction
title_short Continual Learning for Deep Dense Prediction
title_full Continual Learning for Deep Dense Prediction
title_fullStr Continual Learning for Deep Dense Prediction
title_full_unstemmed Continual Learning for Deep Dense Prediction
title_sort continual learning for deep dense prediction
publisher Virginia Tech
publishDate 2018
url http://hdl.handle.net/10919/83513
work_keys_str_mv AT lokegaonkarsanketavinash continuallearningfordeepdenseprediction
_version_ 1719356468051836928