내용 |
Foundations of machine learning -- Using AdaBoost to minimize training error -- Direct bounds on the generalization error -- The margins explanation for boosting's effectiveness -- Game theory, online learning, and boosting -- Loss minimization and generalizations of boosting -- Boosting, convex optimization, and information geometry -- Using confidence-rated weak predictions -- Multiclass classification problems -- Learning to rank -- Attaining the best possible accuracy -- Optimally efficient boosting -- Boosting in continuous time.
|