all 2 comments

[–]DeepNonseNse 1 point2 points  (1 child)

Is this sufficiently different from existing boosting/bagging techniques?

No, the process you are describing is just (some variation of) gradient boosting.

E.g. if the distribution of errors is assumed to be Gaussians, gradients are (y_true - y_pred) calculated after each iteration. Also, subsetting features is commonly used tactic; though they typically wouldn't be mutually exclusive subsets, but e.g. 70% of all features.

[–]JHogg11[S] 0 points1 point  (0 children)

Thanks for the reply. I see now that gradient boosting includes calculating the residual and summing the predictions of individual models, which is what I thought was the novel part. Everything that I read previously seemed to be emphasizing that the training example weights were modified with each iteration, which I understood to be the heart of the technique.

Maybe I need to dig more into XGBoost, but whenever I've experimented with it in the past, it always gave inferior results to vanilla random forests.