
What if 88% of your AI queries didn't need a massive data center, but could run directly on your laptop?
In this episode, we dive into "Intelligence per Watt"—a new metric redefining how we measure AI efficiency. We explore how smaller, local models are rapidly catching up to frontier giants, potentially saving billions in energy costs and democratizing access to intelligence.
Inspired by the work of Jon Saad-Falcon, Avanika Narayan, and their team at Stanford and Together AI, this episode was created using Google’s NotebookLM.
Read the original paper here: https://arxiv.org/abs/2511.07885v1