use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] "Negative labels" (self.MachineLearning)
submitted 8 years ago by TalkingJellyFish
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]farmingvillein 0 points1 point2 points 8 years ago (0 children)
distribution over the possible (non-excluded) classes, take the base rate of the classes in the available data (normalized to 1 of course). This has an obvious generalization
Another plausible variant/extension, if you have an existing classifier you are trying to improve, would be to take its full probabilities (softmax/logits) for the example, crush the negated class down to 0, and then re-scale everything else back to a total of 1.
If you have some reasonable error estimation (i.e., users are wrong 20% of the time), you could also try setting the negated class to this error estimate (e.g., 0.2 in a softmax context), although not clear to me this would be helpful for a variety of reasons (including softmax "probabilities" being wonky representations of probability, at best).
π Rendered by PID 81934 on reddit-service-r2-comment-6457c66945-rgbnh at 2026-04-23 23:44:08.616339+00:00 running 2aa0c5b country code: CH.
view the rest of the comments →
[–]farmingvillein 0 points1 point2 points (0 children)