Visible to Intel only — GUID: GUID-A85CA0DE-63C2-4C17-B6B8-C8D692368C0D
Visible to Intel only — GUID: GUID-A85CA0DE-63C2-4C17-B6B8-C8D692368C0D
Boosting
Boosting is a set of algorithms intended to build a strong classifier from an ensemble of weighted weak learners by iterative re-weighting according to some accuracy measure for weak learners. A weak learner is a classification or regression algorithm that has only slightly better performance than random guessing. Weak learners are usually very simple and fast, and they focus on classification of very specific features.
Boosting algorithms include LogitBoost, BrownBoost, AdaBoost, and others. A Decision Stump classifier is one of the popular weak learners.
In oneDAL, a weak learner is:
Classification algorithm for AdaBoost and BrownBoost
Regression algorithm for LogitBoost
Weak learners support training of the boosting model for weighted datasets.
oneDAL boosting algorithms pass pointers to weak learner training and prediction objects through the parameters of boosting algorithms. Use the getNumberOfWeakLearners() method to determine the number of weak learners trained.
You can implement your own weak learners by deriving from the appropriate interface classes:
Classification for AdaBoost and BrownBoost
Regression for LogitBoost
The number from in case of binary classification.
Class label from for nClasses > 2.
Some boosting algorithms like SAMME.R AdaBoost that require probabilities of classes. For description of each boosting algorithm, refer to a corresponding section in this document.