😺 🎙️ We are LIVE Now

PLUS: WTF is "AI inference"? Our new interview w/ SambaNova Systems explains it!
October 9, 2025
In Partnership with

Welcome, humans.

🔴 We are LIVE decoding Artificial Analysis! Time to find out which AI model actually deserves your money... which provider is secretly 5x cheaper... and what those confusing benchmark charts are really telling us!

Click here to join the live!

Ever wonder if you're getting ripped off paying full price for your AI? Or whether that “SOTA” benchmark score actually means anything for your use case?

We're doing something that's been criminally overdue: a full walkthrough of Artificial Analysis—the site that tracks every major AI model's performance, pricing, and speed across providers. Grant and Corey are going full data detective mode, and honestly? You're about to find out you've been overpaying this whole time.

Why are we talking about this? AMD just signed a multi-billion dollar deal with OpenAI, sending their stock soaring 28% in a single day. Meanwhile, OpenAI announced an even larger $100 billion partnership with NVIDIA and Oracle secured a jaw-dropping $300 billion agreement for AI infrastructure.

Everyone's scrambling to solve the same problem: AI inference is expensive AF.

When you type a prompt into ChatGPT, you're actually triggering a massively expensive computation that's getting exponentially worse.

As Kwasi Ankomah from SambaNova Systems reveals in our latest podcast: "Agents use a phenomenal amount of tokens"10 to 20 times more than regular chatbots.

This hidden bottleneck is about to become everyone's problem, and Kwasi shows how his company’s revolutionary chip architecture already delivers 700+ tokens per second on 90% less power than traditional hardware.

AI Inference: Why Speed Matters More Than You Think

Here’s some of our favorite moments:

The bottom line: While OpenAI, NVIDIA, and Oracle are throwing hundreds of billions at traditional infrastructure, companies like SambaNova are quietly revolutionizing the physics of AI computation. Their "model swapping" capability lets you run cheap models for simple tasks and expensive ones for complex reasoning—all on the same hardware, cutting costs by 40-50x.

This isn't just about faster chatbots. It's about whether AI agents will actually be economically viable at scale.

Listen / Watch now on YouTube | Spotify | Apple Podcasts

Dive deeper with these resources:

Stay curious,

The Neuron Team

That’s all for today, for more AI treats, check out our website.

ICYMI: check out our most recent episodes below!

What do you think of these new podcast episodes?

Pick an option below and share why in the "additional feedback" comment box.

P.P.S: Love the newsletter, but don’t want these new podcast announcement emails? Don’t unsubscribe — adjust your preferences to opt out of them here instead.

cat carticature

See you cool cats on X!

Get your brand in front of 500,000+ professionals here
www.theneuron.ai/newsletter/we-are-live-now

Get the latest AI

email graphics

right in

email inbox graphics

Your Inbox

Join 450,000+ professionals from top companies like Disney, Apple and Tesla. 100% Free.