Continual Learning for Deep Dense Prediction
Transferring a deep learning model from old tasks to a new one is known to suffer from the catastrophic forgetting effects. Such forgetting mechanism is problematic as it does not allow us to accumulate knowledge sequentially and requires retaining and retraining on all the training data. Existing t...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Published: |
Virginia Tech
2018
|
Subjects: | |
Online Access: | http://hdl.handle.net/10919/83513 |
id |
ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-83513 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-VTETD-oai-vtechworks.lib.vt.edu-10919-835132020-11-12T05:42:55Z Continual Learning for Deep Dense Prediction Lokegaonkar, Sanket Avinash Computer Science Ramakrishnan, Naren Huang, Jia-Bin Huang, Bert Computer Vision Continual Learning Image Segmentation Dense Prediction Transferring a deep learning model from old tasks to a new one is known to suffer from the catastrophic forgetting effects. Such forgetting mechanism is problematic as it does not allow us to accumulate knowledge sequentially and requires retaining and retraining on all the training data. Existing techniques for mitigating the abrupt performance degradation on previously trained tasks are mainly studied in the context of image classification. In this work, we present a simple method to alleviate catastrophic forgetting for pixel-wise dense labeling problems. We build upon the regularization technique using knowledge distillation to minimize the discrepancy between the posterior distribution of pixel class labels for old tasks predicted from 1) the original and 2) the updated networks. This technique, however, might fail in circumstances where the source and target distribution differ significantly. To handle the above scenario, we further propose an improvement to the distillation based approach by adding adaptive l2-regularization depending upon the per-parameter importance to the older tasks. We train our model on FCN8s, but our training can be generalized to stronger models like DeepLab, PSPNet, etc. Through extensive evaluation and comparisons, we show that our technique can incrementally train dense prediction models for novel object classes, different visual domains, and different visual tasks. Master of Science 2018-06-12T08:00:31Z 2018-06-12T08:00:31Z 2018-06-11 Thesis vt_gsexam:16094 http://hdl.handle.net/10919/83513 In Copyright http://rightsstatements.org/vocab/InC/1.0/ ETD application/pdf Virginia Tech |
collection |
NDLTD |
format |
Others
|
sources |
NDLTD |
topic |
Computer Vision Continual Learning Image Segmentation Dense Prediction |
spellingShingle |
Computer Vision Continual Learning Image Segmentation Dense Prediction Lokegaonkar, Sanket Avinash Continual Learning for Deep Dense Prediction |
description |
Transferring a deep learning model from old tasks to a new one is known to suffer from the catastrophic forgetting effects. Such forgetting mechanism is problematic as it does not allow us to accumulate knowledge sequentially and requires retaining and retraining on all the training data. Existing techniques for mitigating the abrupt performance degradation on previously trained tasks are mainly studied in the context of image classification. In this work, we present a simple method to alleviate catastrophic forgetting for pixel-wise dense labeling problems. We build upon the regularization technique using knowledge distillation to minimize the discrepancy between the posterior distribution of pixel class labels for old tasks predicted from 1) the original and 2) the updated networks. This technique, however, might fail in circumstances where the source and target distribution differ significantly. To handle the above scenario, we further propose an improvement to the distillation based approach by adding adaptive l2-regularization depending upon the per-parameter importance to the older tasks. We train our model on FCN8s, but our training can be generalized to stronger models like DeepLab, PSPNet, etc. Through extensive evaluation and comparisons, we show that our technique can incrementally train dense prediction models for novel object classes, different visual domains, and different visual tasks. === Master of Science |
author2 |
Computer Science |
author_facet |
Computer Science Lokegaonkar, Sanket Avinash |
author |
Lokegaonkar, Sanket Avinash |
author_sort |
Lokegaonkar, Sanket Avinash |
title |
Continual Learning for Deep Dense Prediction |
title_short |
Continual Learning for Deep Dense Prediction |
title_full |
Continual Learning for Deep Dense Prediction |
title_fullStr |
Continual Learning for Deep Dense Prediction |
title_full_unstemmed |
Continual Learning for Deep Dense Prediction |
title_sort |
continual learning for deep dense prediction |
publisher |
Virginia Tech |
publishDate |
2018 |
url |
http://hdl.handle.net/10919/83513 |
work_keys_str_mv |
AT lokegaonkarsanketavinash continuallearningfordeepdenseprediction |
_version_ |
1719356468051836928 |