Mixture of Experts with Entropic Regularization for Data Classification
datacite.alternateIdentifier.citation | ENTROPY,Vol.21,,2019 | |
datacite.alternateIdentifier.doi | 10.3390/e21020190 | |
datacite.creator | Peralta Márquez, Billy | |
datacite.creator | Saavedra, Ariel | |
datacite.creator | Caro Saldivia, Luis | |
datacite.creator | Soto, Alvaro | |
datacite.date | 2019 | |
datacite.subject.english | mixture-of-experts | |
datacite.subject.english | regularization | |
datacite.subject.english | entropy | |
datacite.subject.english | classification | |
datacite.title | Mixture of Experts with Entropic Regularization for Data Classification | |
dc.date.accessioned | 2021-04-30T16:47:50Z | |
dc.date.available | 2021-04-30T16:47:50Z | |
dc.description.abstract | Today, there is growing interest in the automatic classification of a variety of tasks, such as weather forecasting, product recommendations, intrusion detection, and people recognition. Mixture-of-experts is a well-known classification technique; it is a probabilistic model consisting of local expert classifiers weighted by a gate network that is typically based on softmax functions, combined with learnable complex patterns in data. In this scheme, one data point is influenced by only one expert; as a result, the training process can be misguided in real datasets for which complex data need to be explained by multiple experts. In this work, we propose a variant of the regular mixture-of-experts model. In the proposed model, the cost classification is penalized by the Shannon entropy of the gating network in order to avoid a winner-takes-all output for the gating network. Experiments show the advantage of our approach using several real datasets, with improvements in mean accuracy of 3-6% in some datasets. In future work, we plan to embed feature selection into this model. | |
dc.identifier.uri | http://repositoriodigital.uct.cl/handle/10925/3580 | |
dc.language.iso | en | |
dc.publisher | MDPI | |
dc.source | ENTROPY | |
oaire.resourceType | Article | |
uct.catalogador | WOS | |
uct.indizacion | SCI |