Abstract
In recent years, advances in pattern recognition and data analytics have spurred the development of a plethora of machine learning algorithms and tools. However, as each algorithm exhibits different behavior for different types of data, one is motivated to judiciously fuse multiple algorithms in order to find the “best” performing one, for a given dataset. Ensemble learning aims to create such a high-performance meta-learner, by combining the outputs from multiple algorithms. The present work introduces a simple blind scheme for learning from ensembles of classifiers. Blind refers to the combiner who has no knowledge of the ground-truth labels that each classifier has been trained on. While most current works presume that all classifiers are independent, this work introduces a scheme that can handle dependencies between classifiers. Preliminary tests on synthetic data showcase the potential of the proposed approach.
Original language | English (US) |
---|---|
Title of host publication | 2018 26th European Signal Processing Conference, EUSIPCO 2018 |
Publisher | European Signal Processing Conference, EUSIPCO |
Pages | 2025-2029 |
Number of pages | 5 |
ISBN (Electronic) | 9789082797015 |
DOIs | |
State | Published - Nov 29 2018 |
Event | 26th European Signal Processing Conference, EUSIPCO 2018 - Rome, Italy Duration: Sep 3 2018 → Sep 7 2018 |
Publication series
Name | European Signal Processing Conference |
---|---|
Volume | 2018-September |
ISSN (Print) | 2219-5491 |
Other
Other | 26th European Signal Processing Conference, EUSIPCO 2018 |
---|---|
Country/Territory | Italy |
City | Rome |
Period | 9/3/18 → 9/7/18 |
Bibliographical note
Funding Information:Work in this paper was supported by NSF grants 1500713 and 1514056. Emails: {traga003,georgios}@umn.edu 1The terms annotator, learner, and classifier will be used interchangeably.
Publisher Copyright:
© EURASIP 2018.
Keywords
- Dependent classifiers
- Ensemble learning
- Multi-class classification
- Unsupervised