Intel Rips, Thrive Launches Eternal, GPT 5.5 | Diet TBPN
This episode examines why Intel stock surged 20% after hours following earnings that revealed a fundamental shift in AI infrastructure demand—from GPU-centric training to CPU-intensive agent workflows that require a 1:4 CPU-to-GPU ratio (up from 1:8). The hosts break down five converging demand tailwinds (government backing, Elon Musk's Terrafab project, hyperscaler expansion, and the rise of agentic AI) that position Intel as a surprising beneficiary of the AI boom, despite years of underperformance, and discuss implications for builders: compute supply constraints are real, inference token prices are rising, and the infrastructure layer is consolidating around a few winners.
Key takeaways
- • AI agents require significantly more CPU capacity than GPU-centric training models, shifting the CPU-to-GPU ratio from 1:8 to 1:4 (or potentially 8:1), creating immediate supply constraints and pricing power for chip manufacturers and inference providers.
- • Compute scarcity is structural, not cyclical: even tier-2 and tier-3 AI labs will be sold out of tokens due to infrastructure limitations, meaning margins will expand for model labs and hardware suppliers until supply catches up—a multi-year dynamic.
- • The US government's 10% stake in Intel is creating a national security mandate for domestic chip manufacturing that subsidizes Intel's foundry expansion and guarantees demand from hyperscalers, reducing execution risk and enabling ambitious fab buildouts like Terrafab.
- • Anthropic and OpenAI's real revenue comes from Fortune 500 companies and mainstream users paying directly, not VC subsidies, meaning AI adoption is self-sustaining and the "circular economy" concern is overblown—actual value creation is broad-based across 25+ categories.
- • GPT 5.5 shows qualitatively better reasoning on casual/colloquial prompts (e.g., "Car, bro, it's a car wash" response to a logic question) and correctly solving previously difficult tasks like strawberry-counting, suggesting real capability improvements over prior versions.
- • Builders should prepare for sustained token pricing power: demand for inference vastly exceeds current supply, so expect LLM token prices to remain sticky or rise in the near term, making compute efficiency and prompt optimization increasingly valuable competitive edges.
More from these creators
Gary Vee Says the Smartest Investors Are Going Analog
Intel Rips, Cursor's Plan, Thrive's Giant Bet, GPT 5.5
Meet the New Thiel Fellows
SpaceX-Cursor Deal, ChatGPT Images 2.0, Fake Bear Scam | Diet TBPN
Cursor? The Space Company, Craziest Images 2.0, Spirit Bailout
Apple’s New CEO John Ternus, Ferrari’s First EV | Diet TBPN