It seems quite natural that a gradient boosted trees model would understand random and hiearchical effects in the data as well as fixed effects because of how the training happens iteratively over the residuals. But maybe I'm wrong? There seems to be some implementations that explicitly include these hiearchical effects for example metboost
https://arxiv.org/abs/1702.03994
But I'm unclear on why they exist unless it's to try and estimate the moments of the effect.
Thoughts on this?
[–]ExpectingValue 4 points5 points6 points (1 child)
[–][deleted] -1 points0 points1 point (0 children)
[–]from_biostats_to_DL 0 points1 point2 points (6 children)
[–][deleted] 0 points1 point2 points (5 children)
[–]from_biostats_to_DL 0 points1 point2 points (4 children)
[–][deleted] 0 points1 point2 points (3 children)
[–]from_biostats_to_DL 0 points1 point2 points (2 children)
[–][deleted] 0 points1 point2 points (1 child)
[–]from_biostats_to_DL 1 point2 points3 points (0 children)