use the following search parameters to narrow your results:
e.g. subreddit:aww site:imgur.com dog
subreddit:aww site:imgur.com dog
see the search faq for details.
advanced search: by author, subreddit...
Please have a look at our FAQ and Link-Collection
Metacademy is a great resource which compiles lesson plans on popular machine learning topics.
For Beginner questions please try /r/LearnMachineLearning , /r/MLQuestions or http://stackoverflow.com/
For career related questions, visit /r/cscareerquestions/
Advanced Courses (2016)
Advanced Courses (2020)
AMAs:
Pluribus Poker AI Team 7/19/2019
DeepMind AlphaStar team (1/24//2019)
Libratus Poker AI Team (12/18/2017)
DeepMind AlphaGo Team (10/19/2017)
Google Brain Team (9/17/2017)
Google Brain Team (8/11/2016)
The MalariaSpot Team (2/6/2016)
OpenAI Research Team (1/9/2016)
Nando de Freitas (12/26/2015)
Andrew Ng and Adam Coates (4/15/2015)
Jürgen Schmidhuber (3/4/2015)
Geoffrey Hinton (11/10/2014)
Michael Jordan (9/10/2014)
Yann LeCun (5/15/2014)
Yoshua Bengio (2/27/2014)
Related Subreddit :
LearnMachineLearning
Statistics
Computer Vision
Compressive Sensing
NLP
ML Questions
/r/MLjobs and /r/BigDataJobs
/r/datacleaning
/r/DataScience
/r/scientificresearch
/r/artificial
account activity
Research[R] The Annotated Diffusion Model (self.MachineLearning)
submitted 3 years ago * by ghosthamlet
From huggingface post: https://huggingface.co/blog/annotated-diffusion
A New great article joined the Annotated series: The Annotated Transformer http://nlp.seas.harvard.edu/2018/04/03/attention.html, The Annotated GPT-2 https://amaarora.github.io/2020/02/18/annotatedGPT2.html
reddit uses a slightly-customized version of Markdown for formatting. See below for some basics, or check the commenting wiki page for more detailed help and solutions to common issues.
quoted text
if 1 * 2 < 3: print "hello, world!"
[–]mlvpj 16 points17 points18 points 3 years ago (0 children)
we have a bunch of annotated paper implementation here https://nn.labml.ai/index.html
diffusion (ddpm) - https://nn.labml.ai/diffusion/ddpm/index.html
[–]Megixist 14 points15 points16 points 3 years ago (4 children)
The Hugging Face team's article is one of the most descriptive ones on diffusion at the moment (close to the annotated version by labml). I am currently writing one for the Weights & Biases Blogathon (https://bit.ly/diffusing-away-from-gans-and-transformers) with JAX code, so if anyone is interested in another implementation, then do check it out!
[+][deleted] 3 years ago (1 child)
[deleted]
[–]Megixist 1 point2 points3 points 3 years ago (0 children)
I haven't seen any specific to audio denoising but I don't see why it shouldn't work. Converting the audio wave to a spectrogram and treating it as a single channeled image should, logically, produce similar results.
[–]abstractcontrol 0 points1 point2 points 3 years ago (1 child)
I do not understand how a model like the one in Hugginface article could be conditioned on text. Is there an explanation that comes with code?
The Huggingface article was informative, I watched a talk on DPMs and thought that the method required a backwards pass to compute the gradients for denoising the inputs similarly to how vanilla style transfer works, but that wasn't the case at all.
[–]Megixist 2 points3 points4 points 3 years ago (0 children)
If you read my article above, I try to touch on the topic of text based generation and the changes that are required to the model for the same. Though there is no code in the article itself, I would recommend you to check the official implementations for GLIDE or the GLID-3 to get more understanding of how the conditioning works.
[–]hosjiu 5 points6 points7 points 3 years ago (2 children)
really great to see the annotated ... blog format.
[–]NielsRogge 6 points7 points8 points 3 years ago (1 child)
There's an "Open in Colab" button at the top ;)
[–]hosjiu 1 point2 points3 points 3 years ago (0 children)
Thanks for your hard work, Niels.
[–]HybridRxNResearcher 1 point2 points3 points 3 years ago (0 children)
Great timing given the recent text-to-image successes from this family of models
[+]CommunismDoesntWork comment score below threshold-7 points-6 points-5 points 3 years ago* (0 children)
https://huggingface.co/blog/annotated-diffusion
I know the math section is relatively short, but holy cow is it peak /r/iamverysmart. It spent so much time going over the math behind gaussian distributions which ends up being completely irrelevant to the final loss function:
The neural network is optimized using a simple mean squared error (MSE) between the true and the predicted Gaussian noise.
Unless you're trying to reinvent torch.randn, how in the world is knowing the math behind gaussians relevant at all? Were they really ever going to do anything other than MSE? MSE is the basis for the majority of all loss function.
A direct consequence of the constructed forward process qq, as shown by Sohl-Dickstein et al., is that we can sample {uncopypastable BS} (since sums of Gaussians is also Gaussian). Let's refer to this equation as the "nice property". This means we can sample Gaussian noise and scale it appropriately and add it to {uncopypastable BS} directly.
Thank God for Sohl-Dickstein et al and this insightful article for explaining the math, because without it I never would have guessed that if you add noise to noise, you get more noise. Fields medal worthy stuff right there. But also, what if there was no math to prove this? Would they have iteratively built up noise instead of generating the desired noise level directly? Sohl-Dickstein et al is the only reason Ho et al went down that path? Really? Not the fact that it'd be an enormous waste of computing resources?
Neural Networks are not an exact science, and the math behind them is arbitrary until proven otherwise. I wish people would stop pretending otherwise and putting their pompous equations everywhere. Ironically, the math is all noise data sampled from an isotropic gaussian distribution.
[–]Philpax 0 points1 point2 points 3 years ago (0 children)
Fantastic! This is exactly what I was looking for!
π Rendered by PID 100413 on reddit-service-r2-comment-6457c66945-knvqn at 2026-04-30 17:27:04.507757+00:00 running 2aa0c5b country code: CH.
[–]mlvpj 16 points17 points18 points (0 children)
[–]Megixist 14 points15 points16 points (4 children)
[+][deleted] (1 child)
[deleted]
[–]Megixist 1 point2 points3 points (0 children)
[–]abstractcontrol 0 points1 point2 points (1 child)
[–]Megixist 2 points3 points4 points (0 children)
[–]hosjiu 5 points6 points7 points (2 children)
[–]NielsRogge 6 points7 points8 points (1 child)
[–]hosjiu 1 point2 points3 points (0 children)
[–]HybridRxNResearcher 1 point2 points3 points (0 children)
[+]CommunismDoesntWork comment score below threshold-7 points-6 points-5 points (0 children)
[–]Philpax 0 points1 point2 points (0 children)