Log in

goodpods headphones icon

To access all our features

Open the Goodpods app
Close icon
Kabir's Tech Dives - Diffusion LLMs: A Paradigm Shift in Text Generation

Diffusion LLMs: A Paradigm Shift in Text Generation

03/08/25 • 9 min

4 Listeners

Kabir's Tech Dives

In a groundbreaking development, Diffusion Large Language Models are revolutionizing the field by generating entire responses at once, using a technique inspired by text-to-image generation. This innovative approach, developed by Inception Labs, promises to be 10 times faster and 10 times less expensive than traditional autoregressive models that generate one token at a time. Unlike autoregressive models, diffusion models refine a rough, almost nonsensical text into a coherent solution through iterative steps. This leap in speed, achieving over a thousand tokens per second on standard NVIDIA H100 chips, drastically reduces waiting times and enables more test time compute. This breakthrough not only accelerates coding processes but also facilitates more advanced reasoning, error correction, and controllable generation, opening new possibilities for AI agents, edge applications, and various use cases. According to AI experts like Andrej Karpathy, this diffusion model may also unlock new unique psychology or new strengths and weaknesses, potentially leading to new behaviors in intelligent models.

Send us a text

Support the show

Podcast:
https://kabir.buzzsprout.com
YouTube:
https://www.youtube.com/@kabirtechdives
Please subscribe and share.

plus icon
bookmark

In a groundbreaking development, Diffusion Large Language Models are revolutionizing the field by generating entire responses at once, using a technique inspired by text-to-image generation. This innovative approach, developed by Inception Labs, promises to be 10 times faster and 10 times less expensive than traditional autoregressive models that generate one token at a time. Unlike autoregressive models, diffusion models refine a rough, almost nonsensical text into a coherent solution through iterative steps. This leap in speed, achieving over a thousand tokens per second on standard NVIDIA H100 chips, drastically reduces waiting times and enables more test time compute. This breakthrough not only accelerates coding processes but also facilitates more advanced reasoning, error correction, and controllable generation, opening new possibilities for AI agents, edge applications, and various use cases. According to AI experts like Andrej Karpathy, this diffusion model may also unlock new unique psychology or new strengths and weaknesses, potentially leading to new behaviors in intelligent models.

Send us a text

Support the show

Podcast:
https://kabir.buzzsprout.com
YouTube:
https://www.youtube.com/@kabirtechdives
Please subscribe and share.

Previous Episode

undefined - 😬 GitHub Copilot: AI Security Breach Exposes Private Repos

😬 GitHub Copilot: AI Security Breach Exposes Private Repos

4 Recommendations

GitHub's Copilot experienced a security breach where it leaked sensitive data from previously public repositories that were later made private. Researchers discovered that Copilot retained information even after the repositories were no longer public, impacting over 16,000 organizations. Microsoft initially classified the issue as low severity, drawing criticism for its handling of user privacy. The AI model could regurgitate sensitive data, like API keys and proprietary code, potentially introducing leaked information into other projects. Experts recommend immediately rotating keys and credentials that were ever in a public repo. The incident highlights the risks of AI models training on public data that later becomes private, a growing concern in AI security.

Send us a text

Support the show

Podcast:
https://kabir.buzzsprout.com
YouTube:
https://www.youtube.com/@kabirtechdives
Please subscribe and share.

Next Episode

undefined - Why DevSecOps is Key for Preventing Cyber Security Risks - WABBI

Why DevSecOps is Key for Preventing Cyber Security Risks - WABBI

4 Recommendations

This is the very first interview of my Founder Interview Series. It took me a long time to get this first interview out. One hundred percent my faults. I took too much time to edit and got distracted by a thousand things that were happening around the end of last year.
My sincere apologies to Brittany Greenfield for the delay. But it is now out here and I am super excited to share this very first interview with you all.
** About Brittany Greenfield **
Brittany Greenfield, Founder & CEO of Wabbi, is redefining how companies integrate security into fast-moving DevOps pipelines. With a background spanning top tech firms like Cisco, an MBA from MIT Sloan, and a deep passion for innovation, she’s on a mission to make security seamless without slowing down development. In this episode of Kabir’s Tech Dives, we explore how Wabbi is revolutionizing application security, the role of cybersecurity in development, and why modern businesses must rethink their approach to risk. Brittany also shares insights from her journey as an award-winning cybersecurity leader and how startups can balance agility with resilience.
Wabbi Website:
https://wabbisoft.com
Brittany Greenfield:
https://linkedin.com/in/brittanygreenfield

Send us a text

Everyday AI: Your daily guide to grown with Generative AI
Can't keep up with AI? We've got you. Everyday AI helps you keep up and get ahead.

Listen on: Apple Podcasts Spotify

Support the show

Podcast:
https://kabir.buzzsprout.com
YouTube:
https://www.youtube.com/@kabirtechdives
Please subscribe and share.

Episode Comments

Generate a badge

Get a badge for your website that links back to this episode

Select type & size
Open dropdown icon
share badge image

<a href="https://goodpods.com/podcasts/kabirs-tech-dives-594483/diffusion-llms-a-paradigm-shift-in-text-generation-86947519"> <img src="https://storage.googleapis.com/goodpods-images-bucket/badges/generic-badge-1.svg" alt="listen to diffusion llms: a paradigm shift in text generation on goodpods" style="width: 225px" /> </a>

Copy