Mixture of Experts with Entropic Regularization for Data Classification

Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. “Mixture-of-experts„ is a well-known classification technique; it is a probabilistic model co...

Full description

Bibliographic Details
Main Authors: Billy Peralta, Ariel Saavedra, Luis Caro, Alvaro Soto
Format: Article
Language:English
Published: MDPI AG 2019-02-01
Series:Entropy
Subjects:
Online Access:https://www.mdpi.com/1099-4300/21/2/190
id doaj-d7adedd2001242c892276ce01f66acdf
record_format Article
spelling doaj-d7adedd2001242c892276ce01f66acdf2020-11-24T20:44:29ZengMDPI AGEntropy1099-43002019-02-0121219010.3390/e21020190e21020190Mixture of Experts with Entropic Regularization for Data ClassificationBilly Peralta0Ariel Saavedra1Luis Caro2Alvaro Soto3Department of Engineering Science, Andres Bello University, Santiago 7500971, ChileDepartment of Engineering Informatics, Catholic University of Temuco, Temuco 4781312, ChileDepartment of Engineering Informatics, Catholic University of Temuco, Temuco 4781312, ChileDepartment of Computer Sciences, Pontifical Catholic University of Chile, Santiago 7820436, ChileToday, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. “Mixture-of-experts„ is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a “winner-takes-all„ output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3⁻6% in some datasets. In future work, we plan to embed feature selection into this model.https://www.mdpi.com/1099-4300/21/2/190mixture-of-expertsregularizationentropyclassification
collection DOAJ
language English
format Article
sources DOAJ
author Billy Peralta
Ariel Saavedra
Luis Caro
Alvaro Soto
spellingShingle Billy Peralta
Ariel Saavedra
Luis Caro
Alvaro Soto
Mixture of Experts with Entropic Regularization for Data Classification
Entropy
mixture-of-experts
regularization
entropy
classification
author_facet Billy Peralta
Ariel Saavedra
Luis Caro
Alvaro Soto
author_sort Billy Peralta
title Mixture of Experts with Entropic Regularization for Data Classification
title_short Mixture of Experts with Entropic Regularization for Data Classification
title_full Mixture of Experts with Entropic Regularization for Data Classification
title_fullStr Mixture of Experts with Entropic Regularization for Data Classification
title_full_unstemmed Mixture of Experts with Entropic Regularization for Data Classification
title_sort mixture of experts with entropic regularization for data classification
publisher MDPI AG
series Entropy
issn 1099-4300
publishDate 2019-02-01
description Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. “Mixture-of-experts„ is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a “winner-takes-all„ output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3⁻6% in some datasets. In future work, we plan to embed feature selection into this model.
topic mixture-of-experts
regularization
entropy
classification
url https://www.mdpi.com/1099-4300/21/2/190
work_keys_str_mv AT billyperalta mixtureofexpertswithentropicregularizationfordataclassification
AT arielsaavedra mixtureofexpertswithentropicregularizationfordataclassification
AT luiscaro mixtureofexpertswithentropicregularizationfordataclassification
AT alvarosoto mixtureofexpertswithentropicregularizationfordataclassification
_version_ 1716817345076264960