BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION
Supervised machine learning needs high quality, densely sampled and labelled training data. Transfer learning (TL) techniques have been devised to reduce this dependency by adapting classifiers trained on different, but related, (source) training data to new (target) data sets. A problem in TL is...
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Copernicus Publications
2017-05-01
|
Series: | ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences |
Online Access: | http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/IV-1-W1/229/2017/isprs-annals-IV-1-W1-229-2017.pdf |
id |
doaj-cb88d6f6676a40f5b23a2a629bfe2a42 |
---|---|
record_format |
Article |
spelling |
doaj-cb88d6f6676a40f5b23a2a629bfe2a422020-11-25T02:24:48ZengCopernicus PublicationsISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences2194-90422194-90502017-05-01IV-1-W122923610.5194/isprs-annals-IV-1-W1-229-2017BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATIONK. Vogt0A. Paul1J. Ostermann2F. Rottensteiner3C. Heipke4Institut für Informationsverarbeitung, Leibniz Universität Hannover, GermanyInstitute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, GermanyInstitut für Informationsverarbeitung, Leibniz Universität Hannover, GermanyInstitute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, GermanyInstitute of Photogrammetry and GeoInformation, Leibniz Universität Hannover, GermanySupervised machine learning needs high quality, densely sampled and labelled training data. Transfer learning (TL) techniques have been devised to reduce this dependency by adapting classifiers trained on different, but related, (source) training data to new (target) data sets. A problem in TL is how to quantify the relatedness of a source quickly and robustly, because transferring knowledge from unrelated data can degrade the performance of a classifier. In this paper, we propose a method that can select a nearly optimal source from a large number of candidate sources. This operation depends only on the marginal probability distributions of the data, thus allowing the use of the often abundant unlabelled data. We extend this method to multi-source selection by optimizing a weighted combination of sources. The source weights are computed using a very fast boosting-like optimization scheme. The run-time complexity of our method scales linearly in regard to the number of candidate sources and the size of the training set and is thus applicable to very large data sets. We also propose a modification of an existing TL algorithm to handle multiple weighted training sets. Our method is evaluated on five survey regions. The experiments show that our source selection method is effective in discriminating between related and unrelated sources, almost always generating results within 3% in overall accuracy of a classifier based on fully labelled training data. We also show that using the selected source as training data for a TL method will additionally result in a performance improvement.http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/IV-1-W1/229/2017/isprs-annals-IV-1-W1-229-2017.pdf |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
K. Vogt A. Paul J. Ostermann F. Rottensteiner C. Heipke |
spellingShingle |
K. Vogt A. Paul J. Ostermann F. Rottensteiner C. Heipke BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences |
author_facet |
K. Vogt A. Paul J. Ostermann F. Rottensteiner C. Heipke |
author_sort |
K. Vogt |
title |
BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION |
title_short |
BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION |
title_full |
BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION |
title_fullStr |
BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION |
title_full_unstemmed |
BOOSTED UNSUPERVISED MULTI-SOURCE SELECTION FOR DOMAIN ADAPTATION |
title_sort |
boosted unsupervised multi-source selection for domain adaptation |
publisher |
Copernicus Publications |
series |
ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences |
issn |
2194-9042 2194-9050 |
publishDate |
2017-05-01 |
description |
Supervised machine learning needs high quality, densely sampled and labelled training data. Transfer learning (TL) techniques have
been devised to reduce this dependency by adapting classifiers trained on different, but related, (source) training data to new (target) data
sets. A problem in TL is how to quantify the relatedness of a source quickly and robustly, because transferring knowledge from unrelated
data can degrade the performance of a classifier. In this paper, we propose a method that can select a nearly optimal source from a large
number of candidate sources. This operation depends only on the marginal probability distributions of the data, thus allowing the use of
the often abundant unlabelled data. We extend this method to multi-source selection by optimizing a weighted combination of sources.
The source weights are computed using a very fast boosting-like optimization scheme. The run-time complexity of our method scales
linearly in regard to the number of candidate sources and the size of the training set and is thus applicable to very large data sets. We
also propose a modification of an existing TL algorithm to handle multiple weighted training sets. Our method is evaluated on five
survey regions. The experiments show that our source selection method is effective in discriminating between related and unrelated
sources, almost always generating results within 3% in overall accuracy of a classifier based on fully labelled training data. We also
show that using the selected source as training data for a TL method will additionally result in a performance improvement. |
url |
http://www.isprs-ann-photogramm-remote-sens-spatial-inf-sci.net/IV-1-W1/229/2017/isprs-annals-IV-1-W1-229-2017.pdf |
work_keys_str_mv |
AT kvogt boostedunsupervisedmultisourceselectionfordomainadaptation AT apaul boostedunsupervisedmultisourceselectionfordomainadaptation AT jostermann boostedunsupervisedmultisourceselectionfordomainadaptation AT frottensteiner boostedunsupervisedmultisourceselectionfordomainadaptation AT cheipke boostedunsupervisedmultisourceselectionfordomainadaptation |
_version_ |
1724853492748648448 |