ID: math/0405356

Complexities of convex combinations and bounding the generalization error in classification

May 18, 2004

View on ArXiv
Vladimir Koltchinskii, Dmitry Panchenko
Mathematics
Probability

We introduce and study several measures of complexity of functions from the convex hull of a given base class. These complexity measures take into account the sparsity of the weights of a convex combination as well as certain clustering properties of the base functions involved in it. We prove new upper confidence bounds on the generalization error of ensemble (voting) classification algorithms that utilize the new complexity measures along with the empirical distributions of classification margins, providing a better explanation of generalization performance of large margin classification methods.

Similar papers 1

Bounding the generalization error of convex combinations of classifiers: balancing the dimensionality and the margins

May 18, 2004

91% Match
Vladimir Koltchinskii, Dmitry Panchenko, Fernando Lozano
Probability
Statistics Theory
Statistics Theory

A problem of bounding the generalization error of a classifier f in H, where H is a "base" class of functions (classifiers), is considered. This problem frequently occurs in computer learning, where efficient algorithms of combining simple classifiers into a complex one (such as boosting and bagging) have attracted a lot of attention. Using Talagrand's concentration inequalities for empirical processes, we obtain new sharper bounds on the generalization error of combined clas...

Find SimilarView on arXiv

Data-dependent Generalization Bounds for Multi-class Classification

June 29, 2017

88% Match
Yunwen Lei, Urun Dogan, ... , Kloft Marius
Machine Learning

In this paper, we study data-dependent generalization error bounds exhibiting a mild dependency on the number of classes, making them suitable for multi-class learning with a large number of label classes. The bounds generally hold for empirical multi-class risk minimization algorithms using an arbitrary norm as regularizer. Key to our analysis are new structural results for multi-class Gaussian complexities and empirical $\ell_\infty$-norm covering numbers, which exploit the...

Find SimilarView on arXiv

Greedy Convex Ensemble

October 9, 2019

87% Match
Tan Nguyen, Nan Ye, Peter L. Bartlett
Machine Learning
Machine Learning

We consider learning a convex combination of basis models, and present some new theoretical and empirical results that demonstrate the effectiveness of a greedy approach. Theoretically, we first consider whether we can use linear, instead of convex, combinations, and obtain generalization results similar to existing ones for learning from a convex hull. We obtain a negative result that even the linear hull of very simple basis functions can have unbounded capacity, and is thu...

Find SimilarView on arXiv

Some Local Measures of Complexity of Convex Hulls and Generalization Bounds

May 18, 2004

87% Match
Olivier Bousquet, Vladimir Koltchinskii, Dmitry Panchenko
Probability

We investigate measures of complexity of function classes based on continuity moduli of Gaussian and Rademacher processes. For Gaussian processes, we obtain bounds on the continuity modulus on the convex hull of a function class in terms of the same quantity for the class itself. We also obtain new bounds on generalization error in terms of localized Rademacher complexities. This allows us to prove new results about generalization performance for convex hulls in terms of char...

Find SimilarView on arXiv

Generalization error for multi-class margin classification

August 27, 2007

87% Match
Xiaotong Shen, Lifeng Wang
Statistics Theory
Statistics Theory

In this article, we study rates of convergence of the generalization error of multi-class margin classifiers. In particular, we develop an upper bound theory quantifying the generalization error of various large margin classifiers. The theory permits a treatment of general margin losses, convex or nonconvex, in presence or absence of a dominating class. Three main results are established. First, for any fixed margin loss, there may be a trade-off between the ideal and actual ...

Find SimilarView on arXiv

Empirical margin distributions and bounding the generalization error of combined classifiers

May 18, 2004

87% Match
Vladimir Koltchinskii, Dmitry Panchenko
Probability

We prove new probabilistic upper bounds on generalization error of complex classifiers that are combinations of simple classifiers. Such combinations could be implemented by neural networks or by voting methods of combining the classifiers, such as boosting and bagging. The bounds are in terms of the empirical distribution of the margin of the combined classifier. They are based on the methods of the theory of Gaussian and empirical processes (comparison inequalities, symmetr...

Find SimilarView on arXiv

On the Current State of Research in Explaining Ensemble Performance Using Margins

June 7, 2019

86% Match
Waldyn Martinez, J. Brian Gray
Machine Learning
Machine Learning
Computation

Empirical evidence shows that ensembles, such as bagging, boosting, random and rotation forests, generally perform better in terms of their generalization error than individual classifiers. To explain this performance, Schapire et al. (1998) developed an upper bound on the generalization error of an ensemble based on the margins of the training data, from which it was concluded that larger margins should lead to lower generalization error, everything else being equal. Many ot...

Find SimilarView on arXiv

Local Rademacher complexities

August 16, 2005

85% Match
Peter L. Bartlett, Olivier Bousquet, Shahar Mendelson
Statistics Theory
Statistics Theory

We propose new bounds on the error of learning algorithms in terms of a data-dependent notion of complexity. The estimates we establish give optimal rates and are based on a local and empirical version of Rademacher averages, in the sense that the Rademacher averages are computed from the data, on a subset of functions with small empirical error. We present some applications to classification and prediction with convex function classes, and with kernel classes in particular.

Find SimilarView on arXiv

Geometrical Complexity of Classification Problems

February 11, 2004

85% Match
Tin Kam Ho
Computer Vision and Pattern ...

Despite encouraging recent progresses in ensemble approaches, classification methods seem to have reached a plateau in development. Further advances depend on a better understanding of geometrical and topological characteristics of point sets in high-dimensional spaces, the preservation of such characteristics under feature transformations and sampling processes, and their interaction with geometrical models used in classifiers. We discuss an attempt to measure such propertie...

Find SimilarView on arXiv

Improved Margin Generalization Bounds for Voting Classifiers

February 23, 2025

85% Match
Mikael Møller Høgsgaard, Kasper Green Larsen
Machine Learning
Data Structures and Algorith...
Statistics Theory
Machine Learning
Statistics Theory

In this paper we establish a new margin-based generalization bound for voting classifiers, refining existing results and yielding tighter generalization guarantees for widely used boosting algorithms such as AdaBoost (Freund and Schapire, 1997). Furthermore, the new margin-based generalization bound enables the derivation of an optimal weak-to-strong learner: a Majority-of-3 large-margin classifiers with an expected error matching the theoretical lower bound. This result prov...

Find SimilarView on arXiv