The hardware and bandwidth for this mirror is donated by dogado GmbH, the Webhosting and Full Service-Cloud Provider. Check out our Wordpress Tutorial.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]dogado.de.
With the help of this package the out of bag learning curve for
random forests can be created for any measure that is available in the
mlr
package.
Supported random forest packages are randomForest
and
ranger
and trained models of these packages with the
train
function of mlr
. Available measures can
be looked up on the mlr tutorial
page.
The main function is OOBCurve
that calculates the
out-of-bag curve depending on the number of trees. With the
OOBCurvePars
function out-of-bag curves can also be
calculated for mtry, sample.fraction and min.node.size for the ranger
package.
Installation:
::install_github("PhilippPro/OOBCurve") devtools
Examples:
library(mlr)
library(ranger)
# Classification
= getTaskData(sonar.task)
data = makeClassifTask(data = data, target = "Class")
sonar.task = makeLearner("classif.ranger", keep.inbag = TRUE, par.vals = list(num.trees = 100))
lrn = train(lrn, sonar.task)
mod
# Alternatively use ranger directly
# mod = ranger(Class ~., data = data, num.trees = 100, keep.inbag = TRUE)
# Alternatively use randomForest
# mod = randomForest(Class ~., data = data, ntree = 100, keep.inbag = TRUE)
# Application of the main function
= OOBCurve(mod, measures = list(mmce, auc, brier), task = sonar.task, data = data)
results # Plot the generated results
plot(results$mmce, type = "l", ylab = "oob-mmce", xlab = "ntrees")
plot(results$auc, type = "l", ylab = "oob-auc", xlab = "ntrees")
plot(results$brier, type = "l", ylab = "oob-brier-score", xlab = "ntrees")
# Regression
= getTaskData(bh.task)
data = makeRegrTask(data = data, target = "medv")
bh.task = makeLearner("regr.ranger", keep.inbag = TRUE, par.vals = list(num.trees = 100))
lrn = train(lrn, bh.task)
mod
# Application of the main function
= OOBCurve(mod, measures = list(mse, mae, rsq), task = bh.task, data = data)
results # Plot the generated results
plot(results$mse, type = "l", ylab = "oob-mse", xlab = "ntrees")
plot(results$mae, type = "l", ylab = "oob-mae", xlab = "ntrees")
plot(results$rsq, type = "l", ylab = "oob-mae", xlab = "ntrees")
# Use OOBCurvePars for OOBCurve of other hyperparameters
library(mlr)
= sonar.task
task
= makeLearner("classif.ranger", predict.type = "prob", num.trees = 1000)
lrn = OOBCurvePars(lrn, task, measures = list(auc))
results plot(results$par.vals, results$performances$auc, type = "l", xlab = "mtry", ylab = "auc")
= makeLearner("classif.ranger", predict.type = "prob", num.trees = 1000, replace = FALSE)
lrn = OOBCurvePars(lrn, task, pars = "sample.fraction", measures = list(mmce))
results plot(results$par.vals, results$performances$mmce, type = "l", xlab = "sample.fraction", ylab = "mmce")
= OOBCurvePars(lrn, task, pars = "min.node.size", measures = list(mmce))
results plot(results$par.vals, results$performances$mmce, type = "l", xlab = "min.node.size", ylab = "mmce")
These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.
Health stats visible at Monitor.