BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION

Supervised machine learning needs high quality, densely sampled and labelled training data. Transfer learning (TL) techniques have been devised to reduce this dependency by adapting classifiers trained on different, but related, (source) training data to new (target) data sets. A problem in TL is...

Full description

Bibliographic Details
Main Authors: K. Vogt, A. Paul, J. Ostermann, F. Rottensteiner, C. Heipke
Format: Article
Language:English
Published: Copernicus Publications 2017-05-01
Series:ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
Online Access:http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/IV-1-W1/229/2017/isprs-annals-IV-1-W1-229-2017.pdf
id doaj-cb88d6f6676a40f5b23a2a629bfe2a42
record_format Article
spelling doaj-cb88d6f6676a40f5b23a2a629bfe2a422020-11-25T02:24:48ZengCopernicus PublicationsISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences2194-90422194-90502017-05-01IV-1-W122923610.5194/isprs-annals-IV-1-W1-229-2017BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATIONK. Vogt0A. Paul1J. Ostermann2F. Rottensteiner3C. Heipke4Institut für Informationsverarbeitung, Leibniz Universität Hannover, GermanyInstitute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, GermanyInstitut für Informationsverarbeitung, Leibniz Universität Hannover, GermanyInstitute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, GermanyInstitute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, GermanySupervised machine learning needs high quality, densely sampled and labelled training data. Transfer learning (TL) techniques have been devised to reduce this dependency by adapting classifiers trained on different, but related, (source) training data to new (target) data sets. A problem in TL is how to quantify the relatedness of a source quickly and robustly, because transferring knowledge from unrelated data can degrade the performance of a classifier. In this paper, we propose a method that can select a nearly optimal source from a large number of candidate sources. This operation depends only on the marginal probability distributions of the data, thus allowing the use of the often abundant unlabelled data. We extend this method to multi-source selection by optimizing a weighted combination of sources. The source weights are computed using a very fast boosting-like optimization scheme. The run-time complexity of our method scales linearly in regard to the number of candidate sources and the size of the training set and is thus applicable to very large data sets. We also propose a modification of an existing TL algorithm to handle multiple weighted training sets. Our method is evaluated on five survey regions. The experiments show that our source selection method is effective in discriminating between related and unrelated sources, almost always generating results within 3% in overall accuracy of a classifier based on fully labelled training data. We also show that using the selected source as training data for a TL method will additionally result in a performance improvement.http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/IV-1-W1/229/2017/isprs-annals-IV-1-W1-229-2017.pdf
collection DOAJ
language English
format Article
sources DOAJ
author K. Vogt
A. Paul
J. Ostermann
F. Rottensteiner
C. Heipke
spellingShingle K. Vogt
A. Paul
J. Ostermann
F. Rottensteiner
C. Heipke
BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
author_facet K. Vogt
A. Paul
J. Ostermann
F. Rottensteiner
C. Heipke
author_sort K. Vogt
title BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
title_short BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
title_full BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
title_fullStr BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
title_full_unstemmed BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
title_sort boosted unsupervised multi-source selection for domain adaptation
publisher Copernicus Publications
series ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
issn 2194-9042
2194-9050
publishDate 2017-05-01
description Supervised machine learning needs high quality, densely sampled and labelled training data. Transfer learning (TL) techniques have been devised to reduce this dependency by adapting classifiers trained on different, but related, (source) training data to new (target) data sets. A problem in TL is how to quantify the relatedness of a source quickly and robustly, because transferring knowledge from unrelated data can degrade the performance of a classifier. In this paper, we propose a method that can select a nearly optimal source from a large number of candidate sources. This operation depends only on the marginal probability distributions of the data, thus allowing the use of the often abundant unlabelled data. We extend this method to multi-source selection by optimizing a weighted combination of sources. The source weights are computed using a very fast boosting-like optimization scheme. The run-time complexity of our method scales linearly in regard to the number of candidate sources and the size of the training set and is thus applicable to very large data sets. We also propose a modification of an existing TL algorithm to handle multiple weighted training sets. Our method is evaluated on five survey regions. The experiments show that our source selection method is effective in discriminating between related and unrelated sources, almost always generating results within 3% in overall accuracy of a classifier based on fully labelled training data. We also show that using the selected source as training data for a TL method will additionally result in a performance improvement.
url http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/IV-1-W1/229/2017/isprs-annals-IV-1-W1-229-2017.pdf
work_keys_str_mv AT kvogt boostedunsupervisedmultisourceselectionfordomainadaptation
AT apaul boostedunsupervisedmultisourceselectionfordomainadaptation
AT jostermann boostedunsupervisedmultisourceselectionfordomainadaptation
AT frottensteiner boostedunsupervisedmultisourceselectionfordomainadaptation
AT cheipke boostedunsupervisedmultisourceselectionfordomainadaptation
_version_ 1724853492748648448