Majority vote learning in PAC-bayesian theory: state of the art and novelty
Paul Viallard
Laboratoire Hubert Curien, Data Intelligence Team, Saint-Étienne
https://paulviallard.github.io/
Date(s) : 18/06/2021 iCal
14h30 - 15h30
In machine learning, ensemble methods are ubiquitous: Boosting, Bagging, Support Vector Machine or Random Forest are famous examples. Here we focus on models expressed as a weighted majority vote. The objective is then to learn a majority vote where its performance is guaranteed on new unseen data. Such guarantee can be estimated with PAC (Probably Approximately Correct) guarantees, a.k.a. generalization bounds, that is obtained by upper-bounding the risk that the majority vote makes an error (through the 0-1 loss). One statistical machine learning theory to provide such bounds in the context of majority votes is the PAC-Bayesian framework. The PAC-Bayesian framework has the advantage to offer bounds that can be optimizable: learning algorithms can be derived. However, a major drawback of this framework is that the classical bounds do not directly provide bounds on the majority vote risk: one has to use a (non-precise) surrogate of the 0-1 loss. In this talk, we recall the state-of-the-ar
https://arxiv.org/abs/2104.13626
Catégories