State of AI in 2026: LLMs, Coding, Scaling Laws, China, Agents, GPUs, AGI | Lex Fridman Podcast #490
Lex Fridman discusses the state of AI in early 2026 with machine learning researchers Sebastian Raschka and Nathan Lambert, examining the competitive landscape between US and Chinese AI labs, the evolution of LLM architectures, and the multiple dimensions of scaling laws. The conversation unpacks how despite fundamental architectural similarities to GPT-2, modern LLMs achieve dramatic capability improvements through advances in post-training, inference scaling, data quality, and systems optimization rather than architectural breakthroughs.
Key takeaways
- • The competitive AI landscape is characterized by resource constraints and organizational execution rather than proprietary technological access, as ideas flow freely between labs through researcher mobility.
- • Scaling laws remain robust across pre-training, post-training, and inference dimensions, though the most attractive gains in 2026 come from inference-time scaling and reinforcement learning rather than simply training larger models.
- • Tool use capabilities—enabling LLMs to call APIs, search the web, and execute code—represent a major unlock that's still underutilized in open-source models and require containerization for safe deployment.
- • Data quality and curation matter more than raw data quantity, with techniques like synthetic data generation, OCR of PDFs, and strategic source mixing proving more efficient than simply scaling token counts.
- • Chinese open-weight models like DeepSeek are gaining adoption not primarily through superior performance but through unrestricted licensing, cost efficiency, and local deployment options, posing strategic challenges to US API-based business models.
- • Mixture of Experts (MoE) architectures and attention mechanism refinements like multi-head latent attention and group query attention enable efficient scaling without proportional compute increases, driving recent open-source model improvements.
Recommendations (27)
"First is Build a Large Language Model from Scratch and Build a Reasoning Model from Scratch. I truly believe in the machine learning world, the best way to learn and understand something is to buil..."
Lex Fridman · ▶ 0:53
"First is Build a Large Language Model from Scratch and Build a Reasoning Model from Scratch. I truly believe in the machine learning world, the best way to learn and understand something is to buil..."
Lex Fridman · ▶ 0:53
"I will regularly have like five Pro queries going simultaneously, each looking for one specific paper or feedback on an equation or something."
Nathan Lambert · ▶ 15:32
"And then for code and any sort of philosophical discussion, I use Claude Opus 4.5. Also always with extended thinking."
Nathan Lambert · ▶ 17:05
"I should say, going to Perplexity here, Sebastian Raschka is a machine learning researcher and author known for several influential books."
Lex Fridman · ▶ 24:05
"I need to get him a copy of Season of the Witch, which is a history of SF from 1960 to 1985, which goes through the hippie revolution, like all the gays taking over the city and that culture emergi..."
Nathan Lambert · ▶ 2:28:05
"I try Claude Code on the web every three to six months, which is just prompting a model to make an update to some GitHub repository that I have"
Sebastian Raschka · ▶ 2:37:55
"The Recursive Language Model paper, that is one of the papers that tries to kind of address the long context thing"
Sebastian Raschka · ▶ 2:46:45
"I should say I use Composer a lot because one of the benefits it has is that it's fast"
Sebastian Raschka · ▶ 3:39:38
"I used that feature before, and I always feel bad because it does that every day, and I rarely check it out"
Sebastian Raschka · ▶ 3:21:02
Mentioned (77)
More from these creators
Rick Beato: Greatest Guitarists of All Time, History & Future of Music | Lex Fridman Podcast #492
Khabib vs Lex: Training with Khabib | FULL EXCLUSIVE FOOTAGE
OpenClaw: The Viral AI Agent that Broke the Internet - Peter Steinberger | Lex Fridman Podcast #491
Paul Rosolie: Uncontacted Tribes in the Amazon Jungle | Lex Fridman Podcast #489
Infinity, Paradoxes, Gödel Incompleteness & the Mathematical Multiverse | Lex Fridman Podcast #488
Deciphering Secrets of Ancient Civilizations, Noah's Ark, and Flood Myths | Lex Fridman Podcast #487