ID: math/0405356

Complexities of convex combinations and bounding the generalization error in classification

May 18, 2004

View on ArXiv

Similar papers 4

The complexity of learning halfspaces using generalized linear methods

November 3, 2012

84% Match
Amit Daniely, Nati Linial, Shai Shalev-Shwartz
Machine Learning
Data Structures and Algorith...

Many popular learning algorithms (E.g. Regression, Fourier-Transform based algorithms, Kernel SVM and Kernel ridge regression) operate by reducing the problem to a convex optimization problem over a vector space of functions. These methods offer the currently best approach to several central problems such as learning half spaces and learning DNF's. In addition they are widely used in numerous application domains. Despite their importance, there are still very few proof techni...

Find SimilarView on arXiv

A Dataset-Level Geometric Framework for Ensemble Classifiers

June 16, 2021

84% Match
Shengli Wu, Weimin Ding
Machine Learning
Artificial Intelligence

Ensemble classifiers have been investigated by many in the artificial intelligence and machine learning community. Majority voting and weighted majority voting are two commonly used combination schemes in ensemble learning. However, understanding of them is incomplete at best, with some properties even misunderstood. In this paper, we present a group of properties of these two schemes formally under a dataset-level geometric framework. Two key factors, every component base cl...

Find SimilarView on arXiv

When does Diversity Help Generalization in Classification Ensembles?

October 30, 2019

84% Match
Yijun Bian, Huanhuan Chen
Machine Learning
Machine Learning

Ensembles, as a widely used and effective technique in the machine learning community, succeed within a key element -- "diversity." The relationship between diversity and generalization, unfortunately, is not entirely understood and remains an open research issue. To reveal the effect of diversity on the generalization of classification ensembles, we investigate three issues on diversity, i.e., the measurement of diversity, the relationship between the proposed diversity and ...

Find SimilarView on arXiv

Less Is More: A Comprehensive Framework for the Number of Components of Ensemble Classifiers

September 9, 2017

84% Match
Hamed Bonab, Fazli Can
Machine Learning
Machine Learning

The number of component classifiers chosen for an ensemble greatly impacts the prediction ability. In this paper, we use a geometric framework for a priori determining the ensemble size, which is applicable to most of existing batch and online ensemble classifiers. There are only a limited number of studies on the ensemble size examining Majority Voting (MV) and Weighted Majority Voting (WMV). Almost all of them are designed for batch-mode, hardly addressing online environmen...

Find SimilarView on arXiv

Optimal Linear Combination of Classifiers

March 1, 2021

84% Match
Georgi Nalbantov, Svetoslav Ivanov
Machine Learning
Artificial Intelligence

The question of whether to use one classifier or a combination of classifiers is a central topic in Machine Learning. We propose here a method for finding an optimal linear combination of classifiers derived from a bias-variance framework for the classification task.

Find SimilarView on arXiv

Estimating a sharp convergence bound for randomized ensembles

March 4, 2013

84% Match
Miles E. Lopes
Probability
Social and Information Netwo...
Statistics Theory
Machine Learning
Statistics Theory

When randomized ensembles such as bagging or random forests are used for binary classification, the prediction error of the ensemble tends to decrease and stabilize as the number of classifiers increases. However, the precise relationship between prediction error and ensemble size is unknown in practice. In the standard case when classifiers are aggregated by majority vote, the present work offers a way to quantify this convergence in terms of "algorithmic variance," i.e. the...

Find SimilarView on arXiv

Multiclass learning with margin: exponential rates with no bias-variance trade-off

February 3, 2022

84% Match
Stefano Vigogna, Giacomo Meanti, ... , Rosasco Lorenzo
Machine Learning
Machine Learning

We study the behavior of error bounds for multiclass classification under suitable margin conditions. For a wide variety of methods we prove that the classification error under a hard-margin condition decreases exponentially fast without any bias-variance trade-off. Different convergence rates can be obtained in correspondence of different margin assumptions. With a self-contained and instructive analysis we are able to generalize known results from the binary to the multicla...

Find SimilarView on arXiv

PAC-Bayesian inductive and transductive learning

May 31, 2006

84% Match
Olivier PMA Catoni
Statistics Theory
Statistics Theory

We present here a PAC-Bayesian point of view on adaptive supervised classification. Using convex analysis, we show how to get local measures of the complexity of the classification model involving the relative entropy of posterior distributions with respect to Gibbs posterior measures. We discuss relative bounds, comparing two classification rules, to show how the margin assumption of Mammen and Tsybakov can be replaced with some empirical measure of the covariance structure ...

Find SimilarView on arXiv

Leveraging PAC-Bayes Theory and Gibbs Distributions for Generalization Bounds with Complexity Measures

February 19, 2024

84% Match
Paul Viallard, Rémi Emonet, Amaury Habrard, ... , Zantedeschi Valentina
Machine Learning
Machine Learning

In statistical learning theory, a generalization bound usually involves a complexity measure imposed by the considered theoretical framework. This limits the scope of such bounds, as other forms of capacity measures or regularizations are used in algorithms. In this paper, we leverage the framework of disintegrated PAC-Bayes bounds to derive a general generalization bound instantiable with arbitrary complexity measures. One trick to prove such a result involves considering a ...

Find SimilarView on arXiv

Confidence sets with expected sizes for Multiclass Classification

August 31, 2016

84% Match
Christophe LAMA Denis, Mohamed LAMA Hebiri
Statistics Theory
Statistics Theory

Multiclass classification problems such as image annotation can involve a large number of classes. In this context, confusion between classes can occur, and single label classification may be misleading. We provide in the present paper a general device that, given an unlabeled dataset and a score function defined as the minimizer of some empirical and convex risk, outputs a set of class labels, instead of a single one. Interestingly, this procedure does not require that the u...

Find SimilarView on arXiv