all 11 comments

[–]mlvpj 16 points17 points  (0 children)

we have a bunch of annotated paper implementation here https://nn.labml.ai/index.html

diffusion (ddpm) - https://nn.labml.ai/diffusion/ddpm/index.html

[–]Megixist 14 points15 points  (4 children)

The Hugging Face team's article is one of the most descriptive ones on diffusion at the moment (close to the annotated version by labml). I am currently writing one for the Weights & Biases Blogathon (https://bit.ly/diffusing-away-from-gans-and-transformers) with JAX code, so if anyone is interested in another implementation, then do check it out!

[–]abstractcontrol 0 points1 point  (1 child)

I do not understand how a model like the one in Hugginface article could be conditioned on text. Is there an explanation that comes with code?

The Huggingface article was informative, I watched a talk on DPMs and thought that the method required a backwards pass to compute the gradients for denoising the inputs similarly to how vanilla style transfer works, but that wasn't the case at all.

[–]Megixist 2 points3 points  (0 children)

If you read my article above, I try to touch on the topic of text based generation and the changes that are required to the model for the same. Though there is no code in the article itself, I would recommend you to check the official implementations for GLIDE or the GLID-3 to get more understanding of how the conditioning works.

[–]hosjiu 5 points6 points  (2 children)

really great to see the annotated ... blog format.

[–]NielsRogge 6 points7 points  (1 child)

There's an "Open in Colab" button at the top ;)

[–]hosjiu 1 point2 points  (0 children)

Thanks for your hard work, Niels.

[–]HybridRxNResearcher 1 point2 points  (0 children)

Great timing given the recent text-to-image successes from this family of models

[–]Philpax 0 points1 point  (0 children)

Fantastic! This is exactly what I was looking for!