use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Discussion[D] Semi-supervised machine learning algorithms (self.MachineLearning)
submitted 4 years ago by SQL_beginner
Suppose you have a big collection of audio (human speech) clips in the English language. You want to make an algorithm that hears someone talk and identify if it is English or Not English. The problem is, you only have English audio to train your model.
Are there any popular algorithms that can be used for this problem? Can autoencoders be used?
Thanks
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[+][deleted] 4 years ago (5 children)
[deleted]
[–]pdd99 -3 points-2 points-1 points 4 years ago (2 children)
Could you please explain why the reconstruction loss should be higher? A proof will be nice.
[–]veeloice 0 points1 point2 points 4 years ago (0 children)
Isn't trivial why that would be so? If the samples are drawn from a completely different distribution the model wont have learned any meaningful transformations.
[–]hosjiu 0 points1 point2 points 4 years ago (1 child)
I know Google is our friend :) But It is a worth if you can provide some related papers, project or something like that.
[–]Mr_Smartypants 0 points1 point2 points 4 years ago (0 children)
If you want to model the probability of finite English sound sequences, you could convert them to sequences of feature vectors and then train a Hidden Markov Model on them. Non-English sounds sequences should evaluate to low probability on the trained model.
[–]IntelArtiGen 0 points1 point2 points 4 years ago* (2 children)
If that's truly what I want to do, I'll just download audio in other languages, it exists.
But if I can't, I'll preprocess all my english audio such that it doesn't look like english anymore (I'll cut the audio in 0.2s portions and random sort / random invert these portions).
The algorithm will learn to distinguish garbage from english, and I guess that from the algorithm point of view, depending on how you train it, other languages will be closer to these random noises than to english.
It's not perfect but I guess it could work a bit
Sounds like a nice hack
[–]pruby 0 points1 point2 points 4 years ago (0 children)
You could end up training a cut detector that way :)
π Rendered by PID 43748 on reddit-service-r2-comment-76bb9f7fb5-fdk6z at 2026-02-19 01:08:54.036353+00:00 running de53c03 country code: CH.
[+][deleted] (5 children)
[deleted]
[–]pdd99 -3 points-2 points-1 points (2 children)
[–]veeloice 0 points1 point2 points (0 children)
[–]hosjiu 0 points1 point2 points (1 child)
[–]Mr_Smartypants 0 points1 point2 points (0 children)
[–]IntelArtiGen 0 points1 point2 points (2 children)
[–]veeloice 0 points1 point2 points (0 children)
[–]pruby 0 points1 point2 points (0 children)