Approaching Concept Drift by Context Feature Partitioning

In this paper we present a new approach to handle concept drift using domain-specific knowledge. More precisely, we capitalize known context features to partition a domain into subdomains featuring static class distributions. Subsequently, we learn separate classifiers for each sub domain and classi...

Full description

Bibliographic Details
Main Authors: Hoffmann, Nico, Kirmse, Matthias, Petersohn, Uwe
Other Authors: Technische Universität Dresden, Fakultät Informatik
Format: Others
Language:English
Published: Saechsische Landesbibliothek- Staats- und Universitaetsbibliothek Dresden 2012
Subjects:
Online Access:http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-83954
http://nbn-resolving.de/urn:nbn:de:bsz:14-qucosa-83954
http://www.qucosa.de/fileadmin/data/qucosa/documents/8395/PCD_Technical_Report.pdf
Description
Summary:In this paper we present a new approach to handle concept drift using domain-specific knowledge. More precisely, we capitalize known context features to partition a domain into subdomains featuring static class distributions. Subsequently, we learn separate classifiers for each sub domain and classify new instances accordingly. To determine the optimal partitioning for a domain we apply a search algorithm aiming to maximize the resulting accuracy. In practical domains like fault detection concept drift often occurs in combination with imbalances data. As this issue gets more important learning models on smaller subdomains we additionally use sampling methods to handle it. Comparative experiments with artificial data sets showed that our approach outperforms a plain SVM regarding different performance measures. Summarized, the partitioning concept drift approach (PCD) is a possible way to handle concept drift in domains where the causing context features are at least partly known.