View source: R/SklearnClassifierSettings.R
setAdaBoost | R Documentation |
Create setting for AdaBoost with python DecisionTreeClassifier base estimator
setAdaBoost( nEstimators = list(10, 50, 200), learningRate = list(1, 0.5, 0.1), algorithm = list("SAMME.R"), seed = sample(1e+06, 1) )
nEstimators |
(list) The maximum number of estimators at which boosting is terminated. In case of perfect fit, the learning procedure is stopped early. |
learningRate |
(list) Weight applied to each classifier at each boosting iteration. A higher learning rate increases the contribution of each classifier. There is a trade-off between the learningRate and nEstimators parameters There is a trade-off between learningRate and nEstimators. |
algorithm |
(list) If ‘SAMME.R’ then use the SAMME.R real boosting algorithm. base_estimator must support calculation of class probabilities. If ‘SAMME’ then use the SAMME discrete boosting algorithm. The SAMME.R algorithm typically converges faster than SAMME, achieving a lower test error with fewer boosting iterations. |
seed |
A seed for the model |
## Not run: model.adaBoost <- setAdaBoost(nEstimators = list(10,50,200), learningRate = list(1, 0.5, 0.1), algorithm = list('SAMME.R'), seed = sample(1000000,1) ) ## End(Not run)
Add the following code to your website.
For more information on customizing the embed code, read Embedding Snippets.