PodParley PodParley
308 - How Image Diffusion Models Work - the 20 minute explainer

EPISODE · Mar 24, 2026 · 24 MIN

308 - How Image Diffusion Models Work - the 20 minute explainer

from Fragmented - AI Developer Podcast · host Kaushik Gopal, Iury Souza

You already know how LLMs work from our popular 20-minute explainer. Now we take it to images. What does Michelangelo have to do with stable diffusion? More than you'd think. Walk away knowing how image generation actually works — and what it has in common with the text models you already understand. Full shownotes at fragmentedpodcast.com. Show Notes Episode 303 - How LLMs work in 20 minutes - text generation VAE -Variational Autoencoder RGB Color model - wikipedia Word2Vec technique - wikipedia Efficient Estimation of Word Representation - original Word2Vec paper by Mikolov et al. High-Resolution Image Synthesis with Latent Diffusion Models - Rombach et al. (2022) — the paper behind Stable Diffusion Image Training data LAION-5B - 5 billion image-text pairs scraped from the web, used to train many image generation models WebLI - Google's internal image-text dataset Michelangelo Get in touch We'd love to hear from you. Email is the best way to reach us or you can check our contact page for other ways. We want to hear all the feedback: what's working, what's not, topics you'd like to hear more on. Contact us Newsletter Youtube Website Co-hosts: Kaushik Gopal Iury Souza [!fyi] We transitioned from Android development to AI starting withEp. #300. Listen to that episode for the full story behind our new direction.

NOW PLAYING

308 - How Image Diffusion Models Work - the 20 minute explainer

0:00 24:44
Play in mini player Transcript not yet generated

No transcript for this episode yet

We transcribe on demand. Request one and we'll notify you when it's ready — usually under 10 minutes.

No similar episodes found.

URL copied to clipboard!