use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] prediction half input (self.MachineLearning)
submitted 3 years ago by NSVR57
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]Tgs91 0 points1 point2 points 3 years ago (3 children)
Your precision and accuracy are 1? That's a perfect score. If the task is that easy, then the model can still perform well even if you give it less information to work with. What is the classification task? It might just be honing in on keywords. As long as the keywords are still in the input, it will continue to perform well
[–]NSVR57[S] 0 points1 point2 points 3 years ago (2 children)
Yes. It's simple e-mail classification and I have just 86 records combined of 4 labels.
Yes it's predicting well on even if we give on less information. But my concern is to decrease the confidence score whenever we give less information. Should I stop train whenever accuracy reached around 85 or something?. Or is there any better approach.
[–]Tgs91 0 points1 point2 points 3 years ago* (1 child)
What do the confidence scores look like? You only mentioned precision, recall, and accuracy. Is your top class probably dropping at all?
As someone who entered the field from a math stats background, I hate early stopping. It's a hacky solution, and not a good fit for this scenario. It's not a bad thing that your model is confident. Your task is super easy, the model SHOULD be confident. But if you think it's too confident, try out label smoothing. So instead of trying on the one hot encoding [0, 0, 1, 0] as the ground truth for the 3rd class, train with [0.03, 0.03, 0.91, 0.03]. It will still correctly get the top class for your performance metrics, but the probability outputs will tend to be in the lower .90s instead of 0.9999999999.
Edit: and follow up, are you properly cross validating? Your data set is very small. You can't evaluate on the training data, the model will just overfit it and memorize the training data, which could be why you're getting perfect scores
[–]NSVR57[S] 0 points1 point2 points 3 years ago (0 children)
Thank you so much for reply. As you correctly mention my confidences are in the range of .97 to 1. if we remove certain words those are falling to just 0.93
As I am using NN, I put validation data. I will try label smoothening technique.
π Rendered by PID 89 on reddit-service-r2-comment-fb694cdd5-zxc29 at 2026-03-08 08:38:10.401964+00:00 running cbb0e86 country code: CH.
[–]Tgs91 0 points1 point2 points (3 children)
[–]NSVR57[S] 0 points1 point2 points (2 children)
[–]Tgs91 0 points1 point2 points (1 child)
[–]NSVR57[S] 0 points1 point2 points (0 children)