Episode #490 from 40:08

Transformers: Evolution of LLMs since 2019

And it may be useful to step back and talk about transformer architecture in general. Yeah, so maybe we should start with GPT-2 architecture, the transformer that was derived from the "Attention Is All You Need" paper.

Why this moment matters

And it may be useful to step back and talk about transformer architecture in general. Yeah, so maybe we should start with GPT-2 architecture, the transformer that was derived from the "Attention Is All You Need" paper.

Starts at 40:08
People and topics
All moments
Transformers: Evolution of LLMs since 2019 chapter timestamp | State of AI in 2026: LLMs, Coding, Scaling Laws, China, Agents, GPUs, AGI | EpisodeIndex