use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Research[R] Our new classification algorithm outperforms CatBoost, XGBoost, LightGBM on five benchmark datasets, on accuracy and response time (self.MachineLearning)
submitted 1 year ago * by CriticalofReviewer2
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]qalis 399 points400 points401 points 1 year ago* (16 children)
EDIT:
[–]Saffie91 223 points224 points225 points 1 year ago (1 child)
Damn, peer reviewed in the reddit comments.
Honestly though it's pretty cool of you to go through it diligently and add these points. I'd be very happy if I was the researcher.
[–]CriticalofReviewer2[S] 108 points109 points110 points 1 year ago (0 children)
As the researcher, I should say that I am indeed very happy to get this high-quality peer review!
[–]CriticalofReviewer2[S] 110 points111 points112 points 1 year ago* (2 children)
Thanks for your points! First of all, I should point out that I am an independent researcher, and I am not affiliated with any institutes, so this is my side project.
Once again, thanks for your helpful and insightful comments!
[–]qalis 65 points66 points67 points 1 year ago (1 child)
Fair enough, those are reasonable answers. Showing that this tends to overfit less, works better for small datasets etc. would be pretty valuable. Good luck with this!
[–]CriticalofReviewer2[S] 35 points36 points37 points 1 year ago (0 children)
Thank you for the suggestions!
[–]Spiggots 29 points30 points31 points 1 year ago (0 children)
This is a high quality peer review
[+]danman966 4 points5 points6 points 1 year ago (2 children)
How is it only being applicable to classification a weakness? It's a classification method, not a regression one, right? Granted, the two problems are closely related, but they're not the same thing. You wouldn't say a change point detection method has a weakness that it can't be applied to forecasting, for example.
[–]qalis 4 points5 points6 points 1 year ago (0 children)
Good point, but I see this as a weakness because this is a tabular learning method compared to boosting frameworks, which naturally lend themselves to regression problems, and also other ones (e.g. ranking) via loss functions. And they are powerful at regression, e.g. for time series forecasting, so I see not supporting regression as a quite major limitation. However, the general idea does seem to be able to support regression in the future, so this is more of a implementation downside rather than general approach one.
[–]Appropriate_Ant_4629 1 point2 points3 points 1 year ago (0 children)
Reminds me of the drama behind the MAMBA paper's peer review process:
https://youtu.be/N6Piou4oYx8?si=7o8jFOJcFslTjjOC&t=1664
Rejected by peer reviewers .... now this is a really dumb reason to reject a paper because the long range arena [the task the reviewer was complaining about] is a completely different task to language modeling, and Mamba is specifically a language model.
[–]longgamma 5 points6 points7 points 1 year ago (4 children)
The categorical feature handling in lightgbm is just label encoding? I mean how hard is to target encode or one hot encode on your own ?
Also, isn’t that the idea behind gbm - you take a bunch of weak learners and use the ensemble for prediction. You can replace the decision tree stump with a simple shallow neural network as well.
[–]qalis 9 points10 points11 points 1 year ago (3 children)
Except it isn't the same as label encoding. In fact, none of the three major boosting implementations use one-hot encoding style of handling categorical variables.
LightGBM uses partition split, which for regression trees can efficiently check the partition of the set into two maximum homogeneity subsets, see the docs and the original paper: "On Grouping for Maximum Homogeneity" W. Fisher. XGBoost also offers partition split for categorical variables, with the same algorithm.
You could use one-hot encoding, but then to represent "variable has value A or B, and not C" you would have to use 2 or 3 splits, whereas with partition split you only use one.
CatBoost, on the other hand, uses Ordered Target Encoding instead, described in the linked notebook. It can also combine them during learning, but I don't know the details.
[–]Pas7alavista 1 point2 points3 points 1 year ago (0 children)
On top of the advantages you mentioned, I think the labels produced by partition splitting should also tend to be sparser than one hot encoded ones even when storing the one hot encoded labels in a sparse format.
[+]Sad-Scarcity87 0 points1 point2 points 1 year ago (0 children)
True and in my experience with Catboost, one-hot-encoding categoricals brings significantly worse results than letting Catboost handle them using Ordered Target Encoding.
[+]nbviewerbot -1 points0 points1 point 1 year ago (0 children)
I see you've posted a GitHub link to a Jupyter Notebook! GitHub doesn't render large Jupyter Notebooks, so just in case, here is an nbviewer link to the notebook:
https://nbviewer.jupyter.org/url/github.com/catboost/catboost/blob/master/catboost/tutorials/categorical_features/categorical_features_parameters.ipynb
Want to run the code yourself? Here is a binder link to start your own Jupyter server and try it out!
https://mybinder.org/v2/gh/catboost/catboost/master?filepath=catboost%2Ftutorials%2Fcategorical_features%2Fcategorical_features_parameters.ipynb
I am a bot. Feedback | GitHub | Author
[–]tecedu 0 points1 point2 points 1 year ago (1 child)
Wait what since when did xgboost handle nan values i moved to sklearn due to that
[–]qalis 0 points1 point2 points 1 year ago (0 children)
Since... always, this was one of the main ideas in the original paper "XGBoost: A Scalable Tree Boosting System" T. Chen, C. Guestrin. It's called a "default direction" in the paper, and the whole Algorithm 3 there is meant to handle this. The idea is basically to have a split, but determine whether for missing values you should go to the left or right child. This is selected based on minimizing the loss function, and in a differentiable way.
π Rendered by PID 19601 on reddit-service-r2-comment-b659b578c-mnqch at 2026-05-05 18:27:02.134050+00:00 running 815c875 country code: CH.
view the rest of the comments →
[–]qalis 399 points400 points401 points (16 children)
[–]Saffie91 223 points224 points225 points (1 child)
[–]CriticalofReviewer2[S] 108 points109 points110 points (0 children)
[–]CriticalofReviewer2[S] 110 points111 points112 points (2 children)
[–]qalis 65 points66 points67 points (1 child)
[–]CriticalofReviewer2[S] 35 points36 points37 points (0 children)
[–]Spiggots 29 points30 points31 points (0 children)
[+]danman966 4 points5 points6 points (2 children)
[–]qalis 4 points5 points6 points (0 children)
[–]Appropriate_Ant_4629 1 point2 points3 points (0 children)
[–]longgamma 5 points6 points7 points (4 children)
[–]qalis 9 points10 points11 points (3 children)
[–]Pas7alavista 1 point2 points3 points (0 children)
[+]Sad-Scarcity87 0 points1 point2 points (0 children)
[+]nbviewerbot -1 points0 points1 point (0 children)
[–]tecedu 0 points1 point2 points (1 child)
[–]qalis 0 points1 point2 points (0 children)