A geometric framework for multiclass ensemble classifiers
نویسندگان
چکیده
Abstract Ensemble classifiers have been investigated by many in the artificial intelligence and machine learning community. Majority voting weighted majority are two commonly used combination schemes ensemble learning. However, understanding of them is incomplete at best, with some properties even misunderstood. In this paper, we present a group these formally under geometric framework. Two key factors, every component base classifier’s performance dissimilarity between each pair evaluated same metric—the Euclidean distance. Consequently, ensembling becomes deterministic problem an can be calculated directly formula. We prove several theorems interest explain their implications for ensembles. particular, compare contrast effect number on types schemes. Some important both discussed. And method to calculate optimal weights presented. Empirical investigation conducted verify theoretical results. believe that results from paper very useful us understand fundamental principles general. The also helpful investigate issues classifiers, such as prediction, diversity, pruning, others.
منابع مشابه
Multiclass Boosting for Weak Classifiers
AdaBoost.M2 is a boosting algorithm designed for multiclass problems with weak base classifiers. The algorithm is designed to minimize a very loose bound on the training error. We propose two alternative boosting algorithms which also minimize bounds on performance measures. These performance measures are not as strongly connected to the expected error as the training error, but the derived bou...
متن کاملClassifier Ensemble Framework: a Diversity Based Approach
Pattern recognition systems are widely used in a host of different fields. Due to some reasons such as lack of knowledge about a method based on which the best classifier is detected for any arbitrary problem, and thanks to significant improvement in accuracy, researchers turn to ensemble methods in almost every task of pattern recognition. Classification as a major task in pattern recognition,...
متن کاملTraining highly multiclass classifiers
Classification problems with thousands or more classes often have a large range of classconfusabilities, and we show that the more-confusable classes add more noise to the empirical loss that is minimized during training. We propose an online solution that reduces the effect of highly confusable classes in training the classifier parameters, and focuses the training on pairs of classes that are...
متن کاملBoosting Nearest Neighbor Classifiers for Multiclass Recognition
This paper introduces an algorithm that uses boosting to learn a distance measure for multiclass k-nearest neighbor classi cation. Given a family of distance measures as input, AdaBoost is used to learn a weighted distance measure, that is a linear combination of the input measures. The proposed method can be seen both as a novel way to learn a distance measure from data, and as a novel way to ...
متن کاملUsing Two-Class Classifiers for Multiclass Classification
The generalization from two-class classification to multiclass classification is not straightforward for discriminants which are not based on density estimation. Simple combining methods use voting, but this has the drawback of inconsequent labelings and ties. More advanced methods map the discriminant outputs to approximate posterior probability estimates and combine these, while other methods...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Machine Learning
سال: 2023
ISSN: ['0885-6125', '1573-0565']
DOI: https://doi.org/10.1007/s10994-023-06406-w