The hardware moat around NVIDIA is cracking. Chris Lattner argues NVIDIA’s dominance is a software lock-in problem, not a silicon one. Its CUDA platform is a 20-year-old system unsuited for modern AI, forcing developers into a proprietary silo.
Google, building TPUs for seven generations, now possesses better scale-out capabilities, according to Lattner. The barrier is Google’s lack of a developer community. Amazon’s Trainium and Inferentia chips are also gaining ground with elite clients like Anthropic.
“Hardware vendors refuse to cooperate, forcing developers into proprietary silos that stifle scaling.”
- Chris Lattner, This Week in AI
This fragmentation is occurring as AI itself dissolves other traditional moats. Ben Horowitz states the old software rule - that you can’t hire your way out of a delay - is dead. A company two years behind can now buy enough GPUs and data to compress development into weeks.
Legacy customer lock-in is evaporating, as AI agents can navigate any interface and migrate data, creating what Horowitz calls a “SaaSpocalypse.” Product lifecycles that once spanned a decade may now last just five weeks.
Private equity is capitalizing on this volatility by targeting a different legacy sector: professional services. The playbook involves buying fragmented accounting or law firms and injecting AI to automate the “bottom 50%” of tasks, collapsing the P&L and bypassing offshoring.
Meanwhile, the physical infrastructure for this AI boom is faltering. Horowitz points to a critical U.S. shortage of electricity, memory, and manufacturing capacity. Servers are shipping without RAM, and the grid cannot support new power-hungry clusters - a hardware crisis that $15 billion in new a16z funding aims to address.
The race is no longer just about chips, but about who can build the complete, functional stack before the lights go out.

