Fragmented - Android Developer Podcast

308 - How Image Diffusion Models Work - the 20 minute explainer

Informações:

Sinopse

You already know how LLMs work from our popular 20-minute explainer. Now we take it to images. What does Michelangelo have to do with stable diffusion? More than you'd think. Walk away knowing how image generation actually works — and what it has in common with the text models you already understand. Full shownotes at fragmentedpodcast.com. Show Notes Episode 303 - How LLMs work in 20 minutes - text generation VAE -Variational Autoencoder RGB Color model - wikipedia Word2Vec technique - wikipedia Efficient Estimation of Word Representation - original Word2Vec paper by Mikolov et al. High-Resolution Image Synthesis with Latent Diffusion Models - Rombach et al. (2022) — the paper behind Stable Diffusion Image Training data LAION-5B - 5 billion image-text pairs scraped from the web, used to train many image generation models WebLI - Google's internal image-text dataset Michelangelo Get in touch We'd love to hear from you. Email is the best way to reach us or you can c