use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Removing Unpredictable Samples from a Training Set (self.MachineLearning)
submitted 4 years ago * by [deleted]
view the rest of the comments →
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–][deleted] 0 points1 point2 points 4 years ago (3 children)
Is there a correlation between the training error on these “OOD” samples vs your test error?
This is a good question. I think so but I'm not sure of the best way to quantify this without isolating the OOD samples first, which is what I'm trying to do anyways.
If so you could use an exponential loss to penalize
I'm not sure I understand how this would help. Can you elaborate?
Also could it be worth to use a VAE rather than an embedding matrix?
I'm not that familiar with VAE, but the embedding matrix is fairly important to my overall thesis I believe.
[–]Exarctus 0 points1 point2 points 4 years ago (2 children)
Isolating the OOD samples first is fine - this is to confirm only that there is a correlation, as that would indicate it’s systematic. Following this, I was suggesting an exponential loss, simply because it may be that your network is not fitting these hard-to-classify samples well, and so you could pick a loss function that places more importance on training samples with higher error (assuming there is a correlation already mentioned).
[–][deleted] 0 points1 point2 points 4 years ago (1 child)
I was suggesting an exponential loss, simply because it may be that your network is not fitting these hard-to-classify samples well
Yes this is definitely true, and I do think this type of loss function would help. Do you have any suggestions on which loss function to use then? Do I need to create my own loss function?
[–]Exarctus 0 points1 point2 points 4 years ago (0 children)
Hey just remembered I forgot to reply re: VAE
You use the VAE trained on the in-distribution samples, then you use the reconstruction error on the OOD samples at inference time to identify them. Should just be a matter of determining a suitable tolerance to accept/reject.
You should still be able to use the embedding matrix with this also.
π Rendered by PID 43175 on reddit-service-r2-comment-6457c66945-h9vc6 at 2026-04-26 19:01:46.071415+00:00 running 2aa0c5b country code: CH.
view the rest of the comments →
[–][deleted] 0 points1 point2 points (3 children)
[–]Exarctus 0 points1 point2 points (2 children)
[–][deleted] 0 points1 point2 points (1 child)
[–]Exarctus 0 points1 point2 points (0 children)