GBDT = Gradient Boosting
+ Decision Tree
another names:
- MART(Multiple Additive Regression Tree)
- GBRT(Gradient Boosting Regression Tree)
- Tree Net
Ensemble learning
why ensemble?
- statistics
- calculation
- representation
ensemble type
bagging
Bootstrap AGGregatING
train weak learner parallelly
low variance
boosting
train weak learner iteratively
low bias
- adaboosting
- gradient boosting
Gradient boosting
optimization in parameter space
optimization goal:
solution for the parameters in the form:
where
steepest-descent
define the increments
- current gradient
is computed
where - linear search learning rate
- the step is taken to be
optimization in function space
we consider
we take solution to be
where
current gradient:
and
the multiplier
and the steepest-descent:
finite data
above approach breaks down when the joint distribution of (y, x) is represented by a finite data sample
and then
train gradient model by:
get learning rate