case class BoostingStrategy(treeStrategy: Strategy, loss: Loss, numIterations: Int = 100, learningRate: Double = 0.1, validationTol: Double = 0.001) extends Serializable with Product
Configuration options for org.apache.spark.mllib.tree.GradientBoostedTrees.
- treeStrategy
Parameters for the tree algorithm. We support regression and binary classification for boosting. Impurity setting will be ignored.
- loss
Loss function used for minimization during gradient boosting.
- numIterations
Number of iterations of boosting. In other words, the number of weak hypotheses used in the final model.
- learningRate
Learning rate for shrinking the contribution of each estimator. The learning rate should be between in the interval (0, 1]
- validationTol
validationTol is a condition which decides iteration termination when runWithValidation is used. The end of iteration is decided based on below logic: If the current loss on the validation set is greater than 0.01, the diff of validation error is compared to relative tolerance which is validationTol * (current loss on the validation set). If the current loss on the validation set is less than or equal to 0.01, the diff of validation error is compared to absolute tolerance which is validationTol * 0.01. Ignored when
org.apache.spark.mllib.tree.GradientBoostedTrees.run()
is used.
- Annotations
- @Since( "1.2.0" )
- Source
- BoostingStrategy.scala
- Alphabetic
- By Inheritance
- BoostingStrategy
- Product
- Equals
- Serializable
- Serializable
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Instance Constructors
-
new
BoostingStrategy(treeStrategy: Strategy, loss: Loss, numIterations: Int = 100, learningRate: Double = 0.1, validationTol: Double = 0.001)
- treeStrategy
Parameters for the tree algorithm. We support regression and binary classification for boosting. Impurity setting will be ignored.
- loss
Loss function used for minimization during gradient boosting.
- numIterations
Number of iterations of boosting. In other words, the number of weak hypotheses used in the final model.
- learningRate
Learning rate for shrinking the contribution of each estimator. The learning rate should be between in the interval (0, 1]
- validationTol
validationTol is a condition which decides iteration termination when runWithValidation is used. The end of iteration is decided based on below logic: If the current loss on the validation set is greater than 0.01, the diff of validation error is compared to relative tolerance which is validationTol * (current loss on the validation set). If the current loss on the validation set is less than or equal to 0.01, the diff of validation error is compared to absolute tolerance which is validationTol * 0.01. Ignored when
org.apache.spark.mllib.tree.GradientBoostedTrees.run()
is used.
- Annotations
- @Since( "1.4.0" )
Value Members
-
def
getLearningRate(): Double
- Annotations
- @Since( "1.2.0" )
-
def
getLoss(): Loss
- Annotations
- @Since( "1.2.0" )
-
def
getNumIterations(): Int
- Annotations
- @Since( "1.2.0" )
-
def
getTreeStrategy(): Strategy
- Annotations
- @Since( "1.2.0" )
-
def
getValidationTol(): Double
- Annotations
- @Since( "1.4.0" )
-
var
learningRate: Double
- Annotations
- @Since( "1.2.0" )
-
var
loss: Loss
- Annotations
- @Since( "1.2.0" )
-
var
numIterations: Int
- Annotations
- @Since( "1.2.0" )
-
def
setLearningRate(arg0: Double): Unit
- Annotations
- @Since( "1.2.0" )
-
def
setLoss(arg0: Loss): Unit
- Annotations
- @Since( "1.2.0" )
-
def
setNumIterations(arg0: Int): Unit
- Annotations
- @Since( "1.2.0" )
-
def
setTreeStrategy(arg0: Strategy): Unit
- Annotations
- @Since( "1.2.0" )
-
def
setValidationTol(arg0: Double): Unit
- Annotations
- @Since( "1.4.0" )
-
var
treeStrategy: Strategy
- Annotations
- @Since( "1.2.0" )
-
var
validationTol: Double
- Annotations
- @Since( "1.4.0" )