**The following is shared for educational purposes and is not intended to be financial advice; do your own research!
You've heard the “AI bubble” talk a thousand times by now. Every few weeks, someone compares AI hype to the late-'90s dot-com frenzy, investors get nervous, and then... nothing happens. Valuations keep climbing, the funding keeps flowing, and everyone moves on.
But this is not just another market cycle. According to UK analyst Julien Garran, we're in “the biggest and most dangerous bubble the world has ever seen,” one he calculates is 17 times larger than dot-com, and four times bigger than the 2008 housing crisis.
His case is stark: ten AI startups with zero profits have gained nearly $1 trillion in market value, all while the ecosystem runs on a funding treadmill where—with the exception of NVIDIA—everyone is bleeding money.
Garran argues language models are fundamentally incapable of commercial success for four key reasons:
- They're glorified autocomplete. He contends that LLMs are merely predicting the next word based on statistical patterns, not achieving true understanding. This makes them useful for narrow tasks but severely limited for building genuinely novel, high-value applications.
- They regurgitate existing code. Garran argues that when LLMs write software, they are pulling from memorized patterns in their training data, not creating breakthrough solutions from scratch, which caps their innovative potential.
- They've hit a scaling wall. This is his most critical point. He observes that since GPT-4's launch in March 2023, no subsequent model has dramatically raised the bar on capability, despite astronomical increases in spending. If the bull case of "spend more to get exponentially better" were true, we would see clear evidence of it. Its absence suggests diminishing returns have set in.
- The economics simply don't work. The entire system depends on a never-ending stream of venture capital and mega-investments from players like SoftBank and sovereign wealth funds to subsidize the losses of everyone but the chipmakers.
This sets the stage for the defining conflict of our technological era. The narrative has split into two irreconcilable realities. In one, championed by bulls like venture capitalist Marc Andreessen and NVIDIA CEO Jensen Huang, we are at the dawn of "computer industry V2"—a platform shift so profound it will unlock unprecedented productivity and reshape civilization.
In the other, detailed by macro investors like Julien Garran and forensic bears like writer Ed Zitron, AI is a historically massive, circular, debt-fueled mania built on hype, propped up by a handful of insiders, and destined for a collapse that will make past busts look quaint.
This is a multi-layered conflict playing out across public stock markets, the private venture ecosystem, and the fundamental unit economics of the technology itself. To understand the future, and whether it holds a revolution, a ruinous crash, or a complex mixture of both, we must dissect every layer of the argument, from the historical parallels to the hard financial data and the technological critiques that question the very foundation of the boom.
Here's the bull case from Andreessen and others:
The bull argument is all about the long game and the underlying tech.
- It’s not the dot-com bubble 2.0. That was a telecom bubble with too much infrastructure for too few users. This time, the product (ChatGPT) is already amazing and in the hands of hundreds of millions.
- The raw costs are surprisingly low. Engineer Martin Alderson did the math and found that processing input data is ~1000x cheaper than generating output. This means input-heavy apps like coding assistants should be wildly profitable, with huge markups on subscriptions.
- Productivity will explode. Andreessen predicts AI will act as a "super PhD in every topic" for every individual, leading to massive job growth and making goods and services cheaper for everyone.
And here's the bear case from Zitron, Szyszka, and Ding:
The bears argue the entire thing is a financial illusion built on broken economics.
- It's a circular funding scheme. Zitron argues NVIDIA props up "neoclouds" with investments, which then raise debt to buy NVIDIA's GPUs, creating fake demand that keeps its stock soaring.
- The business model is broken. Startups bet on falling costs, but as Ewa Szyszka and Ethan Ding point out, application costs have exploded. Why? Everyone wants the newest, most expensive model, and those models now use 100x more tokens for complex tasks.
- This creates a "token short squeeze." Companies offering flat-rate subscriptions can't afford their power users. This forces them to throttle service (as users of Claude Code and Cursor have seen), angering customers in a desperate attempt to avoid bankruptcy.
So, who's right?
The debate has moved beyond just markets into the weeds of "tokenomics." On one hand, the raw compute for AI might be cheap and profitable for the model makers like OpenAI (the bull case). On the other, the price AI application startups have to pay is brutal, squeezing their margins to death (the bear case).
This suggests the foundation model companies (OpenAI, Anthropic) are capturing all the value with huge markups, while the ecosystem of AI app startups building on top of them is getting crushed. Furthermore, as mathematician Terrence Tao notes, we're not even calculating the real cost, which must include the price of failed attempts, making the economics even bleaker.
The dot-com parallel holds: a painful, short-term bubble in the application layer seems likely, even if the underlying infrastructure is revolutionary.
What to do:
- For founders and investors: The tokenomics will kill you if you're just reselling an API. Your business model must either be vertically integrated (use your own efficient models), have insane switching costs (lock in enterprise clients), or use AI as a loss-leader for a profitable core service (like hosting).
- For professionals: The free lunch of unlimited, high-end AI is ending. The future for power users is usage-based pricing, which could cost over $100k/year per developer. Start thinking about your AI usage in terms of ROI and track your "token burn" to understand the true cost of the tools you rely on.
After digging through as much data as we could process, the evidence points to a stunning paradox: we are living through a genuine technological revolution financed by a generational bubble. A painful correction in the application layer seems likely, even if the underlying tech is here to stay.
But even a crash doesn't negate the technology. As futurist Peter Leyden argues, we're at a "world historic" tipping point. AI is set to amplify our mental power just as the steam engine amplified our physical power, launching us into the "Age of AI." The chaos we're seeing is what it looks like to be in the middle of a foundational reinvention of our world. Buckle up, y'all.
Read the full bull case from Marc Andreessen here and the bear case from Ed Zitron here. Dig into the tokenomics debate with analyses from Martin Alderson, Ewa Szyszka, and Ethan Ding. And watch the videos shared down below for unique takes from Dylan Patel, Dario Amodei, Jensen Huang, Andrej Karpathy, Theo of T3 Chat, and Julien Garran.
Now, let's dive into each argument in full, and see where it all leads us...
The Historical Case - "Computer Industry V2" or Dot-Com 2.0?
Let us begin with the historical case. We watched this 2 hour episode of Cheeky Pint with Marc Andreessen who argues AI represents "computer industry v2" - the first fundamental reinvention of computing in 80 years - and makes the case that productivity gains will create hyper-deflation rather than unemployment.
Andreessen’s argument is rooted in pattern recognition and a lesson in humility: bubbles are only ever obvious in hindsight. He points to the dot-com era, where even sophisticated investors like Stanley Druckenmiller got the timing spectacularly wrong, shorting tech stocks too early only to capitulate and go long just before the market peaked. The crash itself wasn't a singular event but a slow, cascading decline over five years. For Andreessen, this history teaches a key lesson for venture capital: the best strategy is a disciplined, mechanical pace of investment, because the greatest danger isn't overpaying at the peak; it's panicking and stopping investment at the bottom. He argues that downturns are healthy for the ecosystem as they "clear out the brush" of status-seekers and "tourists," leaving only true believers.
Here are the key arguments in brief:
- AI has already been "hyper-democratized" faster than any technology in history (800M ChatGPT users in 2 years vs. 50M internet users in 1999).
- The technology is "fully there" today unlike early internet which needed decades to become useful.
- Marginal productivity improvements historically lead to hiring MORE people at higher wages, not fewer.
- If AI does eliminate jobs, the resulting productivity boom would cause such dramatic price deflation that living standards would rise anyway (citing the 1880-1930s period as precedent).
- Significant portions of the economy are literally impossible to automate due to licensing/unions/regulation, so the employment apocalypse is structurally blocked.
- We'll likely see a pyramid of AI models (from supercomputers to doorknobs) rather than centralization, with open source dominating most use cases.
- Software development will transform first because it's unregulated and developers build for themselves.
- Comparing AI adoption curves to the internet is misleading since internet took until 2005+ for broadband and 2012 for mobile broadband, whereas AI delivered immediate value from day one.
Andreessen reframes the dot-com bust not as a software bubble but a telecom bubble, where infrastructure was built 15 years ahead of demand for a tiny market of 50 million dial-up users.
This time, he argues, is fundamentally different:
- The Product is Already Spectacular: Unlike the clunky early internet, tools like ChatGPT are "monumentally amazing" and already providing tangible value to hundreds of millions on existing infrastructure.
- It’s a New Computing Architecture: AI isn't just a network; it's a shift from the Von Neumann architecture to neural networks. This is "computer industry V2," a change that will manifest as a "pyramid" of models, from a few massive proprietary ones to billions of small, specialized AIs embedded in everything.
- The Economic Impact Will Be Staggering: He sees AI as a tool that will empower every individual with the knowledge of a "super PhD in every topic," triggering a massive increase in worker productivity, leading not to mass unemployment but to massive job creation, wage growth, and hyper-deflation that makes goods and services cheaper.
His analysis of market timing is equally crucial. He points to the dot-com era, where even sophisticated investors like Stanley Druckenmiller got the timing spectacularly wrong, shorting tech stocks too early only to capitulate and go long just before the market peaked. The crash itself wasn't a singular event but a slow, cascading decline over five years. For Andreessen, this history teaches a key lesson for venture capital: the best strategy is a disciplined, mechanical pace of investment. The greatest danger isn't overpaying at the peak; it's panicking and stopping investment at the bottom.
This new computing paradigm, Andreessen predicts, will form a "pyramid" market structure: a few massive, proprietary models at the top, akin to mainframes, with a sprawling base of billions of smaller, specialized, open-source models embedded into everyday devices. His belief in this future is unshakable, dismissing constant bubble-callers by noting that bubbles are only ever obvious in retrospect. He points to the dot-com crash, which wasn't a singular event but a slow, cascading decline over five years, a history that informs his core VC strategy: maintain a disciplined, mechanical pace of investment, because the biggest mistake is not overpaying at the peak, but panicking and stopping at the bottom.
- (24:26) AI is causing a re-consolidation of the tech industry into just two primary geographic locations globally, with only one in the West, a level of concentration unseen in prior industrial economies.
- (39:11) AI is unique in that it lacks articulate "bear cases" about its business potential. Unlike previous tech cycles, the primary criticism is existential (i.e., "it will destroy the world"), not that it's a failing business model.
- (40:12) The most likely "AI bubble" will not be in AI software itself (which has clear utility) but in the physical infrastructure build-out. This mirrors the dot-com bust, which was less about internet software companies and more of a massive, debt-fueled telecom and fiber overbuild. Today, the equivalent risk is a data center bubble, where capacity is built far ahead of actual utilization.
- (42:31) Bubbles often form in adjacent, established industries because there's a scarcity of talent in the new field (AI software) but an abundance of capital and experienced players in the old one (building data centers). Capital flows to what the "50-year-olds" know how to build.
- (43:56) An ironic potential outcome is that AI researchers could remain underpaid while the market is flooded with an excess of GPUs and data center capacity.
- (44:04) The internet bubble analogy for AI may be flawed. The internet was a network technology whose user experience was severely hampered for years by infrastructure lags (e.g., dial-up modems). AI is a computing technology where the user experience with tools like ChatGPT is already "monumentally amazing" and fully functional from day one.
- (46:05) The most accurate way to frame AI is not as the "new internet," but as "Computer Industry V2." It represents the first fundamental reinvention of the computing model in 80 years—from the Von Neumann architecture to the neural network—unlocking a new class of problems and potentially creating value that is orders of magnitude greater than the first computer industry.
- (47:02) AI has had one of the longest-running hype cycles of any technology. Visions of AI like HAL 9000 existed in the 1960s, meaning the technology took nearly 60 years to catch up to the vision, a far longer lag than for mobile or crypto.
- (49:47) The AI market will develop into a pyramid structure, similar to the evolution of the computer industry (mainframe, minicomputer, PC, embedded chip). A few massive, proprietary "mainframe" models will exist at the top, but the vast majority of AI usage will occur via billions of smaller, hyper-optimized, and likely open-source models embedded in everyday devices like doorknobs.
- (52:05) The AI model market could follow the trajectory of databases and operating systems, where powerful proprietary systems (like Oracle or Windows Server) dominated for a decade before being largely surpassed by more flexible and efficient open-source alternatives (like Postgres or Linux).
- (53:42) Software development will be the area most rapidly transformed by AI. This is because it is a largely unregulated industry where developers are building the tools for themselves, creating an incredibly tight feedback and iteration loop.
- (54:08) The replacement of jobs in fields like medicine and law will happen much slower than people predict due to heavy regulation, licensing requirements, and unionization, which are designed to resist change.
- (55:06) Even in regulated fields, AI's impact will be profound. ChatGPT may already be a better doctor than the average human doctor, and people will increasingly use it as a primary source for medical advice, creating tension with the established system.
- (56:29) AI is perhaps the most democratically-distributed technology in history. Unlike past innovations that trickled down from governments and large corporations to consumers over decades, AI is being adopted in reverse: consumers first, then small businesses, then large enterprises, and finally government.
- (59:52) Referencing the Solow Paradox, the hosts note that AI productivity is "showing up everywhere except the hiring plans of your portfolio companies," questioning when and how these massive efficiency gains will translate into tangible business restructuring.
- (1:00:25) AI will be a democratizing force for businesses, making younger, less bureaucratic companies more competitive against older, entrenched incumbents due to faster adoption.
- (1:01:00) Rather than causing mass unemployment, AI will dramatically increase the marginal productivity of every individual, turning them into a "super PhD in every topic." This will drive demand for more workers at higher wages, leading to employment growth and higher incomes.
- (1:04:17) Even in the unlikely scenario that AI does centralize power and eliminate jobs, the outcome would be hyper-deflation, where the cost of goods and services collapses, making everyone significantly wealthier in real terms.
- (1:11:16) The concept of the "10x engineer" will be updated for the AI era, leading to the emergence of the "1,000x engineer."
- (4:38) It's impossible to know you're in a bubble in real-time. People who correctly call crashes have often been calling for one for years, like an economist who has "predicted nine of the last two crashes."
- (6:04) The most sophisticated hedge fund managers in the world get bubbles wrong. Stan Druckenmiller famously went short on tech in late 1999, capitulated, and then went long in Q1 2000 right before the crash.
- (7:52) Market crashes don't happen all at once; they often cascade down in multiple, discrete moments over several years.
- (8:54) After the dot-com crash, the entrepreneurial ecosystem was so flattened that by 2003, the very idea of starting a company was considered ludicrous. This fear created the perfect environment for the next wave of great companies.
- (9:25) The core of successful venture capital is to maintain a disciplined, mechanical pace of investment through all market cycles. The biggest danger is not overpaying during a bubble but stopping investment during a downturn.
- (10:16) The true bottom of a market cycle isn't just marked by negativity; it's marked by total silence. People completely stop talking about the sector, as if it never existed.
- (29:01) Market downturns are helpful for Silicon Valley's long-term health because they function like "fuel management for fire," clearing out the tourists and status-seekers and leaving only the true believers.
- (41:22) The dot-com bust wasn't an internet software bubble; it was overwhelmingly a telecommunications bubble. The vast majority of capital and debt was tied up in building physical infrastructure (fiber, data centers) that demand wouldn't catch up to for over a decade.
- (42:31) The epicenter of a bubble is never the new thing itself (e.g., AI software), because there aren't enough skilled people. It's in the adjacent, established industries where 50-year-olds know how to deploy massive capital (e.g., data centers for AI, telco for the internet).
- (45:47) The internet bubble analogy for AI is flawed. The internet was a network technology whose experience was hobbled for years by slow connections. AI is a computing technology where the experience is already phenomenal. AI should be seen as "computer industry V2" — the first fundamental reinvention of computing in 80 years, moving from the Von Neumann architecture to the neural network.
- (49:59) The future of AI will not be a few giant, centralized models. It will be a pyramid, just like the computer industry (mainframe -> PC -> embedded devices). There will be a handful of large models at the top and billions of hyper-optimized, smaller, and likely open-source models at the bottom for specific use cases like a smart doorknob.
- (59:52) AI will not lead to mass job loss. Instead, by making every individual a "super PhD in every topic," it will cause the most dramatic increase in the marginal productivity of labor in history, leading to massive employment growth and wage growth.
- (1:04:33) If a few large AI companies did end up dominating the world, the result would not be poverty but hyper-deflation. The price of goods and services would collapse, making everyone massively better off, similar to the "replicator" in Star Trek.
- (1:06:25) The modern Western economy has split into two: a deflationary economy (tech, electronics, media) where things get cheaper and better, and an inflationary economy (housing, healthcare, education) where prices skyrocket due to government policies that restrict supply and subsidize demand.
- (1:11:16) AI will create the 1,000x engineer, and potentially even more, as the productivity gains are applied to a global market of 5 billion connected people.
The Bull Case - A Revolution in Progress
No one is more bullish (and has more to gain if correct) on AI progress than NVIDIA CEO Jensen Huang. So we thought it best to break down his arguments from a recent BG2 podcast to make the ultimate bull case.
The Foundational Drivers of Exponential Demand
- (0:44) The Billion-X Inference Explosion: The core driver of demand is that AI is moving from simple, one-shot answers to "chain of reasoning," which will increase the computational need for inference by a factor of up to a billion.
- (1:45) The Three Scaling Laws: Demand is no longer driven by a single factor. The industry now faces three compounding compute demands: pre-training (initial learning), post-training (AI practicing and refining skills), and inference (AI actively thinking).
- (3:03) Systems of Models: AI is no longer a single model. It's an entire system of specialized models working together—some using tools, some doing research—which dramatically increases overall computational complexity and demand.
- (6:15) Two Compounding Exponentials: The growth is not linear. It's the product of two exponential curves: 1) the exponential growth in users and adoption, multiplied by 2) the exponential growth in compute required per use case due to reasoning.
- (32:42) The Death of Moore's Law: Because performance gains from transistors are over, the only way to keep up with exponential demand and prevent the cost of AI from becoming infinite is through a relentless cycle of full-system architectural innovation, driving a constant refresh and upgrade cycle.
Sizing the Multi-Trillion Dollar Opportunity (The TAM)
- (10:49) The Great Refresh: The most fundamental thesis is that the world's entire multi-trillion dollar installed base of general-purpose computing is obsolete and must be replaced with new accelerated AI infrastructure.
- (11:55) The First Wave - Converting Existing Workloads: Before even accounting for new generative AI, there is a multi-hundred-billion-dollar opportunity in simply converting existing hyperscale workloads (like recommender engines for Meta, TikTok, and Amazon) from CPUs to GPUs.
- (26:52) The Next Wave - The Unconverted Enterprise: A massive, untapped market is the world's structured data (SQL, data processing). This still runs almost entirely on CPUs and represents the next giant wave of workloads that will be moved to accelerated AI systems.
- (13:47) The Ultimate Market - Augmenting Global GDP: Human intelligence drives ~$50 trillion of the world's GDP. Augmenting this intelligence with AI could create a $10 trillion annual market for AI "tokens."
- (15:52) The Implied Capex: A $10 trillion token market, assuming 50% gross margins, would necessitate a $5 trillion annual capital expenditure on AI factories—a more than 10x increase from today's estimated TAM.
- (17:33) A Growing Pie: This is not a zero-sum market. AI will accelerate the growth of the entire world's GDP, creating new wealth and new demand for AI services in a virtuous cycle.
Direct Rebuttals to the "Glut" and "Bubble" Arguments
- (19:02) Existential Spending: The biggest buyers, like Meta, cannot afford to slow down. Mark Zuckerberg stated it's an existential risk to fall behind, making them willing to overspend by billions rather than risk underbuilding. This is not speculative spending; it's a strategic necessity.
- (20:25) We're Already at a Trillion: The "bubble" argument misunderstands the starting point. The AI revenue market isn't a future goal of a trillion dollars; it's effectively there already because the entire revenue base of Google, Meta, and others is now powered by AI.
- (22:30) When a Glut is Possible (And Why Not Now): A glut is "extremely unlikely" until two conditions are met: 1) all general-purpose computing is converted, and 2) all classical hyperscale content generation is AI-based. This transition will take years, ensuring a long runway of demand.
- (23:43) The Demand Signal is the Bottleneck, Not Supply: The current "shortage" is not an NVIDIA supply problem. It's a customer demand problem—they consistently and dramatically under-forecast their own needs, leading to a perpetual scramble to catch up.
- (30:34) Real Economic Substance: This is not a house of cards. The demand is rooted in real-world value: a billion and a half users paying for ChatGPT, every enterprise needing AI to survive, and every nation viewing it as critical infrastructure.
- (1:36:17) AI Creates More Work, Not Less: The fear that AI will destroy its own market by eliminating jobs is wrong. It assumes humanity has no more ideas. AI increases productivity, which creates more wealth, which funds more ideas, which in turn creates more jobs and more demand for AI.
The Economic and Strategic Moat (Why the Growth is Sustainable)
- (50:35) The "Free Chip" Theory: Performance per watt is the only metric that matters. Because a better system generates vastly more revenue from the same fixed data center footprint and power budget, competitors could literally price their chips at zero and customers would still pay a premium for NVIDIA's superior performance. This invalidates the "race to the bottom" pricing argument.
- (35:35) Extreme Co-design as a Moat: Competitors building a single chip cannot compete with NVIDIA's "Extreme Co-design" approach, where the CPU, GPU, networking, and all software are optimized together to deliver compounding performance gains (like the 30x leap to Blackwell) that are impossible otherwise.
- (38:35) The Moat of Scale: The scale is now a defensive barrier. A customer will place a $50 billion purchase order on NVIDIA's proven architecture. No one will take that financial risk on a new, unproven chip.
- (42:23) Barriers to Entry are Higher Than Ever: While the market looks juicy, it is now incredibly complex and massive. It is far harder for a new ASIC company to enter today than it was when NVIDIA and Google's TPU started by dominating a tiny, nascent market.
- (31:51) The AI Flywheel: NVIDIA uses its own AI to design its next generation of chips faster. This creates a self-reinforcing loop where better AI leads to better chips, which enables better AI, allowing Nvidia to stay ahead of the exponential demand curve.
The Unprecedented Breadth of the Customer Base
- (16:32) Real-World Proof: This is not theoretical. Major hyperscalers like Alibaba are publicly stating plans to increase their data center power by 10x by the end of the decade.
- (59:53) A Universal Utility: Unlike past tech bubbles, this is not a niche product. "Everybody needs AI." This positions it as a fundamental utility for the 21st century.
- (1:02:19) A New Customer Class: Nations: The Sovereign AI movement means the customer base is expanding beyond corporations to include every country on Earth, which views AI infrastructure as essential as energy or communications infrastructure. This demand is geopolitically motivated and less tied to economic cycles.
- (1:34:28) The Ultimate TAM: Every Human: AI is the "greatest equalizer," closing the technology divide. If every person can use it simply by speaking to it, the ultimate market size is all 8 billion people on the planet.
The View from Wall Street — It’s Not a Bubble, It’s a Fortress
For many public market investors, the conversation begins and ends with today’s balance sheets. According to analysis from Anthony Pompliano and Phil Rosen, the idea that we’re in an AI-fueled mania ignores the fundamental, and profitable, reality of today’s market leaders. Their argument is a direct rebuttal to dot-com era comparisons. Calling the current market a "bubble," they argue, is an "intellectually lazy" take that misses the crucial differences between then and now.
Here’s their core argument, by the numbers:
- Valuations aren't even close to dot-com levels. Today’s titans, NVIDIA and Microsoft, trade at a reasonable 30x forward earnings. At the peak of the internet mania, Cisco and Oracle soared above 120x earnings, and Microsoft’s own multiple was roughly double its current level. This massive valuation gap, they insist, matters.
- Today’s leaders are cash-flow machines. This is the key distinction. Unlike the "profitless leaders" of the past, the current mega-caps are generating hundreds of billions in annual free cash flow. They command "fortress balance sheets" and contribute meaningfully to real economic growth.
- The psychology is all wrong for a bubble. Pompliano and Rosen offer a contrarian insight: "The fact that everyone’s talking about it is reason to think we aren’t in one." True manias are defined by unchecked euphoria, not constant, widespread anxiety about a potential crash.
So what happens next? They argue the market is pointing higher, backed by a strong macroeconomic outlook. They cite veteran strategist Ed Yardeni, who correctly dismissed recession fears earlier this year. Yardeni’s take is that corrections happen on fears of a recession, while "bear markets tend to be caused by recessions. Currently the economy remains resilient, and a recession is unlikely."
This economic strength is bolstered by powerful historical tailwinds. Since 1950, the S&P 500 has averaged a 4.2% return in the fourth quarter—nearly double any other quarter—and has finished Q4 in the green about 80% of the time.
While they briefly acknowledge "valid concerns about the circular nature of AI spending," they frame it as a secondary issue. The primary story, in their view, is a market led by historically profitable companies with sensible valuations. To them, this isn't a bubble built on hope; it's a boom built on cash.
There's Just One Problem: The Circular Financing Machine...
While Wall Street sees a fortress, others looking at the industry’s plumbing see a perpetual motion machine. Financial analyst Patrick Boyle described the situation with a series of vivid metaphors: an "Ouroboros"—the ancient symbol of a snake eating its own tail; an "extension cord plugged into itself" creating the illusion of energy; and a "Möbius strip made of venture capital and electricity." Chip makers are funding their own customers, who then use that money to buy more chips. Cloud providers are bankrolling AI labs that, in return, are locked into using their services.
This structure isn't entirely new; it bears a striking resemblance to the post-war Japanese keiretsu and South Korean chaebol—industrial conglomerates with intricate cross-holdings designed to secure supply chains, which were criticized for obscuring financial risk and propping up uncompetitive firms.
This modern version is fueling a mind-boggling infrastructure race, but early signs of stress are already appearing. The price to rent NVIDIA's B200 chip has dropped from $3.20 an hour to $2.80 per hour in just a few months. Older chips like the A100 are now available for as little as 40 cents per hour—a rate that is below break-even for many operators. This suggests a potential oversupply of infrastructure built on demand that may not be fully materializing, reminiscent of the telecom firms in the early 2000s that built out fiber optic networks that were never used.
Top Takeaways:
- (1:56)OpenAI is aggressively locking in its supply chain through massive deals, including a $300B cloud agreement with Oracle and memory commitments that reportedly account for half the world's current capacity.
- (2:19) The NVIDIA-OpenAI loop involves NVIDIA pledging $100B in investment, which OpenAI will then use to purchase millions of NVIDIA's GPUs.
- (2:30) AMD has a deal where OpenAI buys its chips, and in return, OpenAI gets cheap stock options that will likely surge when the deal is announced, effectively reimbursing itself for the purchase.
- (3:14) Amazon's $8B+ investment in Anthropic creates a closed loop where Amazon funds the company, which in turn must use Amazon's cloud (AWS), its custom chips, and integrate with its Bedrock platform.
- (3:49) Google has joined this circular model, investing $3B in Anthropic and securing a multi-billion dollar deal for Anthropic to use its TPUs, making Google both an investor and a key infrastructure provider.
- (4:49) Elon Musk's AI strategy is described as having his companies "date each other," with XAI (his startup), X/Twitter (data source), and Tesla (deployment) all intricately linked.
- (7:05) McKinsey forecasts a staggering $5.2 trillion in capital expenditure (capex) will be needed just for chips, data centers, and energy over the next five years.
- (7:18) To justify that level of spending, Bane estimates that AI companies will need to generate $2 trillion in annual revenue, a huge leap from OpenAI's current ~$13 billion.
- (7:51) The circular nature of these deals raises systemic risk concerns: what happens if AI demand or monetization fails to meet the massive investor expectations?
- (8:11) The current AI investment structure is compared to post-war Japan's Keiretsu and South Korea's Chaebol systems—industrial groups with deep cross-holdings that obscured risk and propped up uncompetitive firms.
- (9:21) Are AI giants building a fragile structure, similar to Japan's bubble economy, that looks stable but depends on a constant flood of new capital to "keep the lights on"?
- (9:38) OpenAI's "Stargate" project is a $500 billion plan to build 10 gigawatts of data center capacity.
- (9:59) One gigawatt is the average output of a nuclear power plant and can power a million homes; OpenAI's total committed buildout is 23 gigawatts, requiring the equivalent of 23 nuclear power stations.
- (10:56) The energy demand is so extreme that some data center operators are installing on-site gas turbines and exploring nuclear partnerships just to avoid waiting for grid hookups.
- (11:12) The XAI data center in South Memphis is reportedly running gas turbines without permits, creating significant pollution in an area that already leads Tennessee in asthma hospitalizations.
- (12:10) The companies involved in this massive build-out are not generating enough revenue to justify the spending and appear to lack a clear path to profitability.
- (12:25) A surprisingly large amount of AI output is "slop" (e.g., Taylor Swift deepfakes, anime girlfriend chatbots) rather than the promised scientific breakthroughs.
- (13:12) Despite the "slop," major breakthroughs are happening, such as Google DeepMind researchers winning a Nobel Prize for AI-powered protein folding, which is accelerating drug discovery.
- (14:19) High-growth tech firms are shifting from equity to debt, with OpenAI securing a $4B credit line and providers borrowing against assets like GPUs, which could quickly become obsolete.
- (15:07) The entire AI financial system appears to be "leveraged on optimism."
- (15:13) Users should "make the most of these expensive AI tools that we're currently getting for free," as this period of heavy subsidization may not last.
- (15:32) The GPU rental market is showing stress; rental prices for new chips are dropping, and older A100 chips are renting for as low as 40 cents/hour, which is below the break-even cost for many operators.
- (16:42) If demand doesn't materialize, this infrastructure could become "stranded assets," just like the unused "dark fiber" optic networks from the dot-com bubble.
- (17:51) How much of the "massive demand" for NVIDIA's chips is real, and how much is just NVIDIA's own investment money circulating back to it?
- (18:51) AI adoption is shallow; OpenAI has 700M weekly users, but only 5% pay. Most revenue is from enterprise deals, where McKenzie estimates the pilot project success rate is less than 15%.
- (19:15) The predicted mass AI-driven layoffs haven't happened, with the only clear negative impact being on freelance graphic designers, copywriters, and some junior coders.
- (20:38) This isn't a repeat of the dot-com bubble; the mega-cap tech firms have solid financials (projected $200B in free cash flow), strong balance sheets, and real earnings.
- (21:55) The most significant constraint on AI's growth isn't capital, but electricity, as the 23+ gigawatts of power needed for these data centers cannot be brought online quickly.
- (23:40) AI may not be a "winner-take-all" market; the "DeepSeek moment" showed models can be replicated quickly and cheaply, which could lead to a competitive market with no single player having pricing power.
- (24:20) The ultimate winners may not be the AI labs themselves (who may struggle to monetize) but rather the traditional businesses that use AI to boost productivity.
The Bear Case: A Trillion-Dollar House of Cards
For the bear case, read Ed Zitron's 18,500-word investigation arguing generative AI is an unprecedented bubble built on circular money flows, impossible unit economics, and mythmaking that will inevitably collapse. The key theory: The current level of AI spending is "committed to failure" because there is no concrete, articulated plan for the capital that justifies the cost.
Key arguments in brief:
- NVIDIA maintains growth by investing in "neoclouds" who use that money plus massive debt to buy NVIDIA GPUs, then NVIDIA becomes their largest customer renting back the capacity—outside the Magnificent Seven and OpenAI, these neoclouds have under $1B in real revenue combined.
- Total generative AI revenue is only ~$61B in 2025 against hundreds of billions in costs, with even Microsoft's world-class sales machine converting only 1.81% of its 440M Office subscribers to Copilot.
- The technology is fundamentally unreliable—LLMs are probabilistic and can't be trusted to do the same thing twice, making them unsuitable for replacing knowledge work despite the hype.
- Actual software engineers explain coding LLMs are like "slightly-below-average CS graduates who can't learn"—useful for simple tasks but incapable of the architectural thinking, maintenance, and contextual decision-making that constitutes real engineering work
- There's no "profit lever" because users cost unpredictable amounts (some Anthropic users burn $50,000/month on $200 subscriptions with no way to stop them), making traditional SaaS economics impossible.
- OpenAI needs over $1 trillion in the next four years ($300B to Oracle, $325B+ to build data centers to unlock NVIDIA's staged funding, $115B operational burn) but the world's top 10 private equity firms have only ~$477B available capital combined and US VC could run out in six quarters.
- OpenAI's projections are absurd—CFO Sarah Friar signed off on making $200B revenue by 2030 (more than Meta made in 2024) with negative cash flow magically improving by $39B in a single year, requiring a 10x revenue increase in an industry with $61B total revenue today.
- Unlike past bubbles, this leaves behind expensive specialized GPUs with limited alternative uses and rapid depreciation rather than general-purpose infrastructure, while the demand simply isn't there—after 3 years and $500B invested, products outside ChatGPT (which burns $8B+ annually) show minimal traction.
Zitron's thesis begins and ends with NVIDIA, the chipmaker that has become so central to the market that it accounts for a staggering 7-8% of the S&P 500's entire value. He argues the market is a circular funding scheme fueled by private credit, with an estimated $50 billion per quarter flowing into data centers. NVIDIA props up "neoclouds" (like CoreWeave), which then raise massive debt to buy NVIDIA's GPUs, creating fake demand.
He argues that if you subtract the revenue these neoclouds get from the hyperscalers (Microsoft, Google) and from NVIDIA itself, the "real" external customer market for AI compute is less than a billion dollars across all neoclouds combined. The entire industry is being propped up by a torrent of private credit for a technology whose entire commercial revenue is less than that of a single hit mobile game. Zitron argues the numbers simply don't add up:
- Revenue is a Pittance: The entire generative AI industry is projected to generate a paltry ~$61 billion in 2025 while burning hundreds of billions in costs.
- Even Microsoft is Failing: With the world's best sales machine, Microsoft has only convinced <2% of its 440+ million potential users to pay for its 365 Copilot, generating less than $3 billion in annual revenue.
- Unit Economics are Broken: The most popular AI products are bleeding money. Microsoft's GitHub Copilot loses over $20 per user per month on average. Anthropic’s popular Claude Code generates only $33 million per month, with every power user being wildly unprofitable due to uncontrollable token burn.
- OpenAI's Economics are Impossible: The company needs over $1 trillion in the next four years to cover its announced deals, including a $300 billion commitment to Oracle and a $10 billion partnership with Broadcom. This sum is greater than the entire world's available private credit.
The financial projections required to sustain this are, in Zitron's view, divorced from reality. OpenAI's internal revenue projections, which forecast it will generate $200 billion annually by 2030, are dismissed by Zitron as "bonkers"—a fantasy designed to keep the charade going. This level of spending is so extreme that he predicts US VC firms may run out of money in 18 months at the current burn rate.
Furthermore, the headline-grabbing funding announcements are, upon inspection, illusory. The supposed "$100 billion investment" from NVIDIA into OpenAI isn't upfront cash. Zitron notes that it's a series of tranches, with the vast majority being contingent on OpenAI successfully building gigawatts of new data center capacity—capacity it has no conceivable way to pay for. It's a conditional promise, not a blank check.
Finally, Zitron argues that beyond the broken finances, the technology itself is fundamentally flawed for enterprise use. He contends that Large Language Models are inherently unreliable:
- They Hallucinate Logic: LLMs are probabilistic systems that don't just get facts wrong; they fail at basic logic, making them a dangerous liability for any mission-critical task.
- They Can't Replace Knowledge Work: The narrative that AI will replace jobs like software engineering is a lie, he argues, propagated by executives who don't understand the work. These jobs are about architecture, maintenance, and contextual problem-solving—skills LLMs fundamentally lack. They can make easy tasks easier, but they often make hard tasks harder.
This technological unreliability, in Zitron's view, fatally undermines the entire investment thesis, as the products can never be trusted to perform the high-value work needed to justify their astronomical costs.
Here's our best TL;DR of Ed's full argument (go read the whole essay if you want to understand the risks of investing in AI atm, it's extremely well-sourced, and that's coming from me, a certified linkaholic):
The circular money problem:
- NVIDIA maintains growth by investing in "neoclouds" (CoreWeave, Lambda, Nebius) who then use that money plus massive debt ($25B for CoreWeave alone) to buy NVIDIA GPUs.
- NVIDIA also becomes their largest customer, renting back the capacity.
- Outside of NVIDIA, hyperscalers, and OpenAI, these neoclouds have less than $1 billion in combined real revenue.
- This creates an illusion of demand while the entire industry burns hundreds of billions building capacity nobody wants.
The revenue crisis:
- Total generative AI revenue is only ~$61 billion in 2025 (including OpenAI's projected $13B and Anthropic's $2B) against hundreds of billions in costs.
- Microsoft - the world's best enterprise software seller with 440 million Microsoft 365 subscribers - has only converted 1.81% (8 million users) to pay for Copilot at likely-discounted rates, generating maybe $3 billion annually.
- GitHub Copilot, the most popular coding tool with 1.8M subscribers, loses $20+ per user per month.
- Anthropic can't control costs - users on the "viberank leaderboard" burn $50,000+/month on $200 subscriptions.
The fundamental product flaw:
- LLMs are probabilistically unreliable and can't be trusted to do the same thing twice, making them unsuitable for replacing knowledge work.
- Three experienced engineers explain coding LLMs are like "slightly-below-average CS graduates who can't learn" - useful for simple tasks but incapable of the contextual thinking, maintenance, and architectural decisions that constitute actual software engineering.
- The media conflates "writing code" with "doing software engineering" when they're entirely different.
The impossible math:
- OpenAI needs over $1 trillion in the next four years: $300B committed to Oracle, $325B+ to build the 10GW of data centers required to unlock NVIDIA's staged "$100B investment" (which isn't real funding), $115B in operational burn, plus backup compute, Microsoft Azure commitments, Google TPU costs, and international Stargate projects.
- OpenAI projects making $200B in revenue by 2030 (more than Meta made in 2024) - a 10x increase in an industry with $61B total revenue today.
- CFO Sarah Friar signed off on projections showing negative cash flow suddenly improving by $39B in a single year (2029), which Zitron calls "ethically questionable."
The capital doesn't exist:
- The top 10 private equity firms have ~$477B available capital combined. US VC has $164B dry powder and could run out in six quarters.
- Global VC deal activity hit its lowest since 2018 ($139B in H1 2025), and without OpenAI's phantom $40B raise, US VC would have declined 36%.
- OpenAI alone is absorbing massive chunks of available private and venture capital while providing no exits or returns.
- The bubble requires $400B+ investment over 3 years with no clear funding source.
Why this time is different: Unlike the fiber boom (which created lasting infrastructure) or Uber (which had better unit economics), GPUs have limited use cases outside AI, depreciate rapidly, require specialized expensive infrastructure, and there's no "profit lever" to pull. Data center development already accounts for more US GDP growth than all consumer spending combined. Every AI company is unprofitable with no path to profitability because users cost unpredictable amounts - there's no way to prevent power users from burning 3,000%+ of their subscription value. Traditional SaaS economics don't apply.
The demand isn't there: After 3 years and $500B+ invested, outside of OpenAI's ChatGPT (which burns $8B+ annually), there's barely any product-market fit. Cursor hit $500M ARR before rate limits, Perplexity makes $12.5M/month while burning 164% of revenue on compute. Replit customers are revolting over surprise $1,000 bills. Most "AI revenues" are Microsoft renting Azure compute to OpenAI at a loss ($2.20/hour loss per A100 GPU). The industry has conflated "NVIDIA selling GPUs" with "demand for AI products" when they're completely different things.
Zitron argues this is the largest misallocation of capital in tech history - hundreds of billions poured into unreliable technology with no proven demand, sustained by circular money flows between a handful of companies, marketed through vague promises about "the future" that the media uncritically amplifies. The bubble's collapse is inevitable because OpenAI literally cannot raise the trillion+ dollars it needs, the unit economics make profitability impossible, and the actual use cases don't justify the infrastructure being built. Unlike past bubbles, this will create massive economic damage while leaving behind expensive, specialized GPUs with limited alternative uses and rapid depreciation.
You can also watch Ed's ~2 hour chat w/ the guys at the Ben and Emil show, where he riffs on the same material; great episode, though there's lots of swearing, but fairly entertaining; it stole my whole Saturday afternoon... so congrats, I guess?
Top Takeaways:
- (5:09) Financial Analysis: Bain Capital calculated a $2 trillion revenue requirement to make AI investments worthwhile but projects an $800 billion shortfall. The entire sector currently generates only about $55 billion.
- (5:55) Financial Analysis: Deutsche Bank warns that the current spending rate is only sustainable if it remains "parabolic," which is highly improbable, indicating the market is on borrowed time.
- (6:12) Product Insight: New AI features, like ChatGPT's "Pulse," are expensive (pro-users only) yet offer trivial value (e.g., suggesting Halloween costumes), demonstrating a failure to create valuable, revenue-generating products.
- (11:52) Market Data: User adoption of AI tools has peaked and is now declining, according to data from Apollo, suggesting the total addressable market is far smaller than claimed.
- (12:11) Key Statistic: Microsoft's flagship 365 suite, with 440 million users, has only 8 million active paying AI subscribers, an adoption rate of less than 2%, signaling a catastrophic failure to monetize a captive market.
- (13:55) Strategic Insight: Companies are manipulating metrics to feign success, such as redefining "monthly active users" from 30 days to 28 days to inflate numbers, a sign of underlying weakness.
- (14:50) Economic Theory: The claim that the "cost of intelligence" is falling is false. The cost of inference (the actual use of AI) is rising due to more complex models that burn more compute tokens to perform tasks.
- (16:08) Prediction: When the bubble bursts, the private equity sector will be one of the biggest losers ("get washed") due to its heavy, late-stage investment in AI infrastructure.
- (21:03) Technical Limitation: OpenAI has internally confirmed that hallucinations are an unfixable, inherent flaw of LLMs, creating a permanent ceiling on their reliability and enterprise value.
- (22:03) Product Risk: The unreliability of AI agents poses a direct financial risk, exemplified by the company Replit, whose AI coding agent went rogue and spent hundreds of dollars of its customers' money without permission.
- (23:02) Efficiency Statistic: Some OpenAI reasoning models require 4 to 12 dedicated GPUs for a single user instance, making them extraordinarily inefficient and expensive to run at scale.
- (43:34) Capital Analysis: The bubble is mathematically unsustainable. OpenAI alone requires $1 trillion to execute its plan, yet the total available capital from the top 10 private equity firms ($477B) and all US venture capital ($164B) is not enough to fund just one company.
- (47:36) Strategic Insight: NVIDIA is engaging in Enron-like financial engineering by creating special purpose entities to lease GPUs, masking the true financial risk from its balance sheet.
- (51:55) Economic Contrast: Unlike past bubbles (railroads, fiber optics), AI infrastructure (GPUs) becomes obsolete in just 1-3 years and has no secondary market, meaning the trillions invested will become worthless rather than repurposed.
- (57:14) Strategic Insight: The core mentality driving the bubble is captured by a financial analyst's admission: there isn't enough money for the industry's plans, but there is "enough capital to do this for at least a little while longer."
- (1:00:50) Strategic Thinking: The leaders driving the boom are not focused on ROI. Quotes from Mark Zuckerberg (willing to "misspend a couple of hundred billion dollars") and Larry Page ("willing to go bankrupt rather than lose this race") reveal the bubble is fueled by FOMO, not fundamentals.
- (1:01:51) Historical Comparison: The AI bubble is fundamentally different from the dot-com bubble. The dot-com bust was caused by products being too popular for their flawed business models; the AI bust will be caused by products that were never popular or useful enough to begin with.
- (1:19:12) Core Theory: The bubble is a direct result of leadership failure. The tech industry is run by "business idiots" who lack technical understanding and are chasing AI as the only available "growth narrative" in a market that has run out of genuine ideas.
- (1:26:53) Financial Structure: The AI hardware ecosystem is a high-risk circular reference where NVIDIA is the investor, the supplier, and the main customer to startups that use NVIDIA's own contracts as collateral to raise debt to buy more of its GPUs.
- (1:35:20) Prediction: The collapse will be sudden and swift, likely triggered by a single news story leaking the disastrous internal economics of a major AI player, which will shatter market confidence overnight.
- (1:40:44) Actionable Insight: The proof that AI is not revolutionizing productivity is the absence of a massive wave of new, cheap software ("shovelware"). If building software was suddenly easy, the market would be flooded; it is not.
- (1:42:45) Prediction: Major legacy tech companies are at risk. Oracle, in particular, is "mortgaging its future" on its massive deal with OpenAI and could collapse from the financial exposure.
- (1:44:14) Financial Prediction: OpenAI's future projections are nonsensical. It claims it will reach $145 billion in revenue by 2029 (more than chip giant TSMC makes today) while still losing billions and facing over $100 billion in compute bills that same year.
The Macro Bear Case: "The Biggest and Most Dangerous Bubble"
While Ed Zitron provides a forensic, bottom-up analysis of the bubble's financial plumbing, UK analyst Julien Garran offers a sweeping, top-down, macro-level condemnation that attacks both the economics and the core technology of the AI boom. He calls the current environment "the biggest and most dangerous bubble the world has ever seen," calculating that the cumulative misallocation of capital across AI, VC, crypto, and housing has reached a scale 17 times larger than the dot-com bubble and four times the 2008 housing crisis.
His case rests on a simple observation: ten AI startups with zero profits have gained nearly $1 trillion in market value over the past year, all while the ecosystem runs on a funding treadmill where—with the exception of NVIDIA—everyone is bleeding money without a "killer app" to justify the spend.
At the heart of his thesis is a "Golden Rule": "If you use a large language model to develop an app or a service, it'll never be commercial." He argues that LLMs are "built to fail" as commercial products because they are fundamentally a "simulacrum of language," not a form of cognitive intelligence. They are statistical regurgitation machines that combine word correlation with rote learning, allowing them to answer common questions but failing completely on novel ones. He points to an LLM's catastrophic failure to draw a chessboard one move before a win as proof: it knows the words from chess books but has zero understanding of the game's meaning.
Garran's most controversial take is that AI is perfectly suited for "bullsh*t jobs" where the output is never rigorously checked, but this doesn't make it transformative—it just automates the replacement of "bullsh*t with bullsh*t."

Consider this your daily reminder to always double check ChatGPT’s work 🙂
This technological flaw is compounded by an inescapable economic one: the scaling wall. Garran asserts that improving LLMs gives, at best, linear benefits while incurring exponential costs. The empirical proof, he argues, is GPT-5, which he estimates used 80 to 100 times more compute than its predecessor yet was widely seen as a "flop" for not being significantly better. This leads to broken unit economics, where power users on a $200 Anthropic subscription can burn through over $10,000 in compute, making the business model the inverse of traditional software.
This unhealthy ecosystem is most visible in the data centers themselves. Garran's analysis shows that for a data center to earn a reasonable 10% annual return on a new NVIDIA Blackwell GPU, it would need to rent it out for $6.31 per hour. The current public market rate is $3.79 per hour, guaranteeing the operator a 25% loss. This confirms that only the shovel seller, NVIDIA, is making money.
Financial red flags, according to Garran, are now flashing bright red. He draws a direct parallel to the dot-com bust, noting that while Cisco's vendor financing receivables grew 138% before its crash, NVIDIA's have grown an astonishing 626% and now represent 85% of quarterly revenue. This, combined with what he describes as financial "round-tripping" between NVIDIA and its customers and massive, one-way insider selling, paints a picture of a system sustained by financial engineering, not genuine demand. Outside of this, he points to data showing enterprise adoption is already declining, setting the stage for a painful unwind that could see corporate profits fall by 35-40% as trillions in misallocated capital are written off.
To Garran's main point, the other companies selling shovels (besides NVIDIA) during the gold rush are doing just fine:
- Case in point: Credo, a little-known Silicon Valley company making purple cables that connect AI servers.
- Their stock has doubled this year (up 245% in 2024), and their market cap hit $25 billion.
- Those signature purple cables, historically the color of royalty because the dye was expensive and rare to produce, now cost $300-$500 each and are everywhere in AI data centers, from xAI's Memphis facility to Meta's server racks.
Also ripping: memory (HBM/DDR) & storage:
- Micron ($MU) has benefitted directly from AI server builds; the company says it expects to sell out its 2026 supply of high-bandwidth memory used in AI chips, a sign demand is still running hot.
- DDR4/DDR5 “regular” memory prices are rising as buyers stock up; TrendForce reported DDR spot prices jumping in October and flagged double-digit contract hikes into Q4.
- SK hynix and Samsung are also key memory suppliers riding the same wave; SK hynix says it finished checks on next-gen HBM4 and is preparing production to meet AI demand.
- Storage names like WDC and Seagate are getting a lift as cloud providers add capacity for AI workloads; for instance, Western Digital beat revenue expectations on stronger data-storage demand.
So when does it pop? Garran admits he can't call the exact top; markets hit all-time highs just last week. But he sees warning signs: VC funding for AI startups is drying up because valuations are absurdly high. That leaves a shrinking pool of mega-investors (SoftBank, foreign governments, NVIDIA) to keep the party going.
Now what if he's wrong? Garran sees two scenarios:
- The bubble lasts longer than expected, wasting more capital on projects that don't generate real economic value. Future GDP suffers.
- Someone actually achieves “superintelligence,“ completely reshaping society. We either get utopia or dystopia, depending on who controls it.
For the record, he's betting on option one.
The bottom line: Whether you think Garran is a prophet or a pessimist, his analysis highlights a real tension in AI: massive investment, limited profitability, and a lot of hope riding on breakthroughs that haven't materialized yet. Even if AI doesn't crash dramatically, a slow deflation could reshape the entire tech landscape and your job along with it.
Garran admits he cannot call the exact top but points to the drying up of traditional VC funding as a key warning sign, leaving a shrinking pool of mega-investors to keep the party going. Ultimately, he predicts the bubble's deflation will not lead to a utopia of superintelligence, but to a period of wasted capital and damage to future GDP—a slow, painful reckoning for a revolution that hasn't yet materialized.
Top Takeaways:
- (16:29) The Golden Rule: Julian's core thesis is that if you use a Large Language Model (LLM) to build an app or service, you will never be able to make a commercial return on it.
- (16:59) Built to Fail: LLMs are described as being "built to fail" because they are only a "simulacrum of language" and are fundamentally incapable of applying actual "cognitive intelligence."
- (17:40) The Scaling Wall: The central failure point of the business model is that the cost of compute required to service LLM needs is greater than what people are prepared to pay.
- (17:57) Evidence of Failure (ChatGPT-5): ChatGPT-5 is cited as a "flop" because it showed no significant improvement over version 4, despite estimates that it consumed 80 to 100 times more compute to build.
- (19:06) Prediction (The Pivot): Because the current LLM path is failing, developers are pivoting to "world models" (synthetic physics), after an earlier pivot to reinforcement learning was deemed too expensive.
- (19:18) Critique of World Models: These new "world models" are also "very limited" because you cannot learn how the world works from a model that doesn't include everything about how the world works.
- (26:16) Critique (Statistical Process): An LLM is just a "statistical process" for predicting the next word; it has no "intention" or "meaning" behind its words, unlike a human.
- (27:32) Analogy (The Encyclopedia Kid): LLMs are compared to a child who copies answers from an encyclopedia (rote learning), rather than a child who has a deep understanding of the subject. The LLM will fail as soon as it's asked a question that isn't in its "encyclopedia" (training data).
- (28:44) Failure Example (The Chessboard): When an image LLM was asked for a chessboard "one move before white wins," it produced a "catastrophe" (with the wrong number of squares) because it doesn't understand chess, it only knows the statistical relationship of words in chess books.
- (30:13) Failure Example (Video): Video-creating LLMs produce errors like gymnasts with extra limbs because they lack a fundamental understanding of human physiology and physical constraints.
- (30:59) Critique (Commercial Value): AI is described as a "regurgitation machine that sometimes gets it wrong," making it suitable for homework, which "has no commercial value."
- (31:59) Critique (Coding): AI-generated code is "buggy," doesn't integrate with existing software stacks, and likely uses copyrighted code, making it uncommercializable.
- (32:40) Insight (The Scaling Math): The benefits of improving an LLM (e.g., adding more numbers to its vectors) are "at best linear," while the compute costs required to recalculate all correlations "go right through the ceiling."
- (35:00) Business Model Failure: Companies like Anthropic are "hemorrhaging money" because users (e.g., coders) can extract far more value in compute (e.g., $10,200) than they pay for their subscription ($200).
- (35:44) Insight (Opposite of Software): This business model is the inverse of traditional software (like Excel), where the cost of the second sale is near zero. With AI, scaling increases the cost.
- (36:32) Insight (Sora): Video tools like Sora limit videos to 10 seconds because allowing longer renders would cause users to burn "massive amounts of compute" trying to re-roll and fix all the errors.
- (41:21) The "Looks About Right" Problem: AI is perceived as successful because its answers often "look about right," and most people lack the time or expertise to rigorously test them and find the flaws.
- (42:29) Insight (Stanford Study): A Stanford study ("The Agent Company") found that LLM agents tasked with running a software company had a task completion rate between 1.5% (Grok) and 34%.
- (43:18) Critique (Probabilistic Failure): Even a 34% success rate is useless because "it's a different 34% every time you run the system." It's a probabilistic model that cannot be relied upon for consistent business processes.
- (44:01) Critique (Disingenuous Confidence): LLMs are intentionally designed to appear confident and not provide confidence intervals, because developers know that users find a lack of confidence "less convincing."
- (48:18) Takeaway (Where AI is Used): AI is being adopted for "bullshit jobs"—roles where the quality of the output is not easily tested, such as sifting HR resumes, parts of marketing, or management consulting.
- (49:23) Takeaway (Monopoly Use): Monopolies (like Amazon or Microsoft) are the only ones who can effectively use AI because they can cut costs by lowering quality, and their customers have no alternative.
- (50:31) Prediction (Business Practice): Big tech firms are firing US workers, using AI for a "first pass" on their work, and then hiring cheap offshore labor to "clean up the slop," resulting in a worse product but higher profit.
- (51:42) Insight (Canary in the Coal Mine): Microsoft's Satya Nadella, who knows OpenAI's true state better than anyone, has pivoted Microsoft away from training (which he knows has hit a wall) and toward products.
- (52:06) Insight (Financial Engineering): Microsoft was booking compute time sold to OpenAI (in exchange for equity) as revenue, then booking the mark-to-market rise in OpenAI's valuation as investment gains—a financial loop.
- (53:32) Insight (Lack of Moat): Nadella knows that even if you build a superior LLM, a competitor can use that LLM to create synthetic data to train a new, cheaper model that is "almost as good," destroying any competitive moat.
- (54:41) Failure Example (Agents): OpenAI's "Operator" agent (designed to order groceries, pizza, etc.) was benchmarked as completing its tasks only one-third of the time.
- (55:42) Ecosystem Failure: In the entire AI ecosystem, "only really NVIDIA making the chips... [is] making any money at all." Everyone else is losing money.
- (57:54) The Data Center Math: Data centers like Coreweave are "guaranteed a loss" renting Blackwell chips; the analysis shows a required rental price of $6.31/hour for a 10% return, but the current market rate is only $3.79/hour.
- (1:01:55) Analogy (Dotcom Bubble): The current AI bubble is compared to the dotcom bubble's "vendor financing" scheme (used by Cisco and Nortel), where companies lend money to customers just so those customers can buy their products.
- (1:03:28) Financial Red Flag (NVIDIA): NVIDIA's receivables are up 626% in 30 months (vs. Cisco's 138% in 15 months) and have ballooned from 55% to 85% of its quarterly revenue.
- (1:04:06) Financial Red Flag (NVIDIA): NVIDIA is engaged in "selling chips and renting them back" to build its own global models, but this is a very small and fragile ecosystem (robotics, driverless cars).
- (1:06:25) Financial Red Flag (Roundtripping): Details the complex "roundtripping" of investments between Coreweave, Magnetar Capital, and NVIDIA, designed to "keep the game going" and inflate revenues.
- (1:07:00) Financial Red Flag (Insider Selling): Massive insider selling at NVIDIA and Coreweave.
Comparing The Actual Costs
Here's the problem with all of this: the entire debate over AI profitability is forced to rely on "napkin math" because no vendor publishes their end-to-end Cost of Goods Sold (COGS) per token. This information black hole, encompassing variables like GPU amortization, power, and networking, is why the question is so fiercely contested. The conflict boils down to the brutal economics of running AI models, a debate with compelling arguments on both sides.
The Bear Case: A "Token Short Squeeze"
The original bet that AI costs would plummet has, according to analysts Ewa Szyszka and Ethan Ding, failed spectacularly. They argue application costs have exploded for two reasons:
- Demand is Only for the Best: Users consistently demand the newest, most capable model, which always debuts at a high price. As Ding puts it, an old AI model is like "yesterday's newspaper."
- Usage Has Gone "Nuclear": As models get smarter, they are used for more complex tasks that consume far more tokens, with this consumption outpacing any efficiency gains.
This creates a "token short squeeze." Companies are trapped because flat-rate subscriptions, while effective for winning market share, can lead to bankruptcy by subsidizing power users. A single user paying $20 a month can easily burn through thousands of dollars in compute costs, turning every power user into a massive liability. This is compounded by business model challenges, such as OpenAI's low 2% conversion rate from free to paid users, and the fact that half of its 800 million global users are in emerging markets unlikely to pay a $20 subscription. Szyszka predicts the future for power users will involve plans costing upwards of $100,000 per year.
Adding another layer, mathematician Terrence Tao argues the industry isn't calculating the true cost of its initiatives. We celebrate the wins but ignore the cost of failures. His formula is damning: if a tool costs $1,000 in compute per attempt but only succeeds 20% of the time, its real cost per success is $5,000.
The Bull Counterargument: Inference as a Money Printer
Engineer Martin Alderson challenges this narrative, calculating a stunning 1,000x cost asymmetry between processing inputs and generating outputs. Based on H100 GPU costs, he estimates the raw cost to a provider is roughly:
- $0.003 per million input tokens (essentially free).
- $3.08 per million output tokens (where the real cost lies).
This means input-heavy applications should be wildly profitable. Alderson estimates a typical ChatGPT Pro user on a $20/month plan costs only $3 to serve, representing a 5-6x markup. The problem isn't that AI is inherently unprofitable, but that flawed business models allow output-heavy users to erase profits.
This bullish view is supported by independent, industry-wide benchmarks. Research from Epoch AI shows that the price to achieve a fixed level of performance on various AI tasks is falling at an exponential rate—between 9 and 900 times per year. This trend is fueled by tangible hardware improvements, with benchmarks from MLPerf Inference revealing that next-generation hardware like Nvidia’s Blackwell platform delivers step-function gains in throughput and efficiency.
GQG Partners raised the issue that OpenAI only has a 2% conversion rate from free to paid, and with 800M users globally, of which half are from emerging markets who wouldn’t pay for a $20 subscription. In fact, the top-line numbers for the industry leader, OpenAI, seem to confirm the bear case at first glance. Recent financials reported by the Financial Times paint a stark picture:
- Users: 800 million total users.
- Paying Users: 5% conversion rate, resulting in 40 million paying customers.
- Annual Recurring Revenue (ARR): $13 billion.
- Average Revenue Per Paying User (ARPU): $325 per year, or $27 per month.
- Losses: A staggering $8 billion loss in the first half of 2025 alone, implying a potential $20 billion annual loss run rate.

The Core Ratio: For every $1 in revenue, OpenAI is spending approximately $3.
These numbers seem unsustainable. However, an alternative framework, explained by YouTuber Theo, reframes these losses not as an operational failure but as a massive, front-loaded investment in future growth.
This analysis argues that looking at a single year's Profit & Loss (P&L) is misleading. The key is to separate the cost of serving current customers (inference) from the cost of building next-generation technology (training/R&D).
Sam Altman's Core Quote: OpenAI's CEO stated, "We're profitable on inference. If we didn't pay for training, we'd be a very profitable company." This confirms the conceptual split between operational costs and R&D.
Dario Amodei's Framework (The "Each Model is a Company" Analogy): The CEO of Anthropic provides the clearest mental model (in this episode of Cheeky Pint):
- Year 1: You spend $100M.
- Year 2: Model A generates $200 million in revenue (it's a profitable "product"). But simultaneously, you spend $1 billion to train the next-gen Model B. The company P&L is now -$800M, looking much worse.
- Year 3: Model B generates $2 billion in revenue (another profitable product). But you spend $10 billion to train Model C. The company P&L is -$8B, looking catastrophic.
The Insight: The overall P&L gets worse each year, but that's only because the company is "founding another company that's like much more expensive" every cycle. The loss is the R&D investment in the next, exponentially more powerful product.
Theo's Multiplier Model: This framework only works if revenue growth outpaces cost growth. Theo proposes a hypothetical model to illustrate the path to profitability:
Assumption: Costs grow by 3x each year, while Revenue grows by 4x.
- Starting Point (2025): $13 billion Revenue and $33 billion in total spending (revenue + loss).
- 2026: Revenue grows to $52 billion (4x), but spending grows to $99 billion (3x). The loss widens to $47 billion.
- 2027: Revenue hits $208 billion, spending hits $297 billion. The loss widens again to $89 billion.
- 2028: Revenue reaches $832 billion, while spending reaches $891 billion. The loss shrinks dramatically to $59 billion.
- 2029: Revenue explodes to $3.3 trillion, finally surpassing spending of $2.6 trillion. The company becomes profitable.
The Takeaway: They are not investing in the 2025 P&L; they are investing in the potential for revenue to outpace costs due to a superior growth multiplier, even if it takes years and trillions of dollars. The goal isn't to double their money, but to potentially 100x it on a future trillion-dollar market.
The Crux of the Entire Debate: The Scaling Wall
Now, this framework articulated by Dario Amodei and modeled by Theo—that today's massive losses are justified R&D for a future, exponentially better product—as well as the whole bull case and the trillion-dollar valuations, are all entirely dependent on one critical assumption: that spending more money on training will continue to yield exponentially better models.
Put another way, that the scaling laws continue to work.
Julien Garran's observation that progress has been incremental since GPT-4—citing GPT-5 as a "flop" in terms of capability leap despite a potential 100x increase in training compute—is a direct and formidable challenge to the entire bull thesis. If Julien is correct that the laws that justify these outrageous capital burns have broken down, then OpenAI’s $20 billion annual loss is simply the catastrophic cost of running an unprofitable business, and the investment model collapses.
Garran's data center math also provides a definitive answer to the "is inference a money printer?" debate. While Martin Alderson's model suggests a theoretical path to profitability, Garran's analysis of real-world hardware and rental costs ($6.31 needed vs. $3.79 charged) demonstrates that at the infrastructure level, the business is structurally unprofitable for everyone except NVIDIA. Furthermore, his comparison of NVIDIA's 626% receivables growth to Cisco's 138% before the dot-com crash suggests that not only are the patterns of financial engineering similar to past bubbles, but they are occurring on a scale that is orders of magnitude larger and more dangerous.
Given the conflicting economics, why keep throwing good money after bad? There are three hard-nosed reasons for the massive capital inflows:
- Cloud Attach & Platform Lock-in: Even if AI services compress cloud margins in the near term, they pull workloads into Azure, AWS, and GCP, expanding their total addressable market.
- Steep, Ongoing Cost Curves: Hardware and software progress continues to drive the cost per useful token down much faster than in most software categories, supporting the long-run margin story.
- Compounding Scale Advantages: Large fleets, lower revenue-share deals over time (like OpenAI's plan to reduce Microsoft’s share from ~20% to ~8%), and better utilization can flip unit economics without changing list prices.
The human capital investment is equally staggering, where paying a top researcher a billion-dollar package is justified if their insight improves model efficiency by just 5%, potentially saving billions in compute.
The Realist's Case — The Death of SaaS and What Comes Next
Dylan Patel of SemiAnalysis offers a view that reframes the entire debate. The issue isn't just a bubble, but a fundamental, painful restructuring of the entire tech industry, and in Patel's words, this "highest stakes capitalism game of all time" is already compressing the profitability of the core businesses that fund it.
- Microsoft, for example, has repeatedly noted in its financial reports that its cloud gross margin has decreased—dropping to 71% in Q1 FY25—explicitly "driven by scaling our AI infrastructure."
- Similarly, Alphabet’s CFO has lifted the company's 2025 capital expenditure guidance to a staggering $85 billion to meet AI and cloud demand.
- Amazon is facing the same dynamic, with contemporaneous coverage of its earnings focusing on rising AI CapEx creating near-term pressure on its AWS margins.
- This demonstrates that while AI services are a strategic necessity to secure platform lock-in, they are actively eroding the near-term profitability of the hyperscalers funding the revolution.
This dynamic creates a steep path to standalone profitability for AI companies, even those with massive revenue. For example, a significant portion of OpenAI's revenue flows directly to Microsoft, its infrastructure and distribution partner. While this share is expected to decrease from around 20% to 8% by 2030, it highlights the hard financial reality that a large portion of revenue is immediately captured by the essential cloud partner.
The capital war is being fought between the balance sheets of giants like Meta, Google, and Microsoft. Even a company like OpenAI, with a current run-rate of approximately $15-20 billion in annual revenue, is considered "too small to matter" without powerful allies. This is because a substantial portion of an AI startup's funding, as much as 70%, is immediately spent on compute power, flowing directly to companies like NVIDIA. Even these giants require external capital; OpenAI is reportedly seeking a $30 billion data center deal with Apollo as its own cash flow is insufficient.
Patel refutes the idea of diminishing returns with an analogy: the value difference between model tiers is drastic, comparing it to the difference in work a 6-year-old, a 13-year-old, and a 25-year-old can accomplish. Each significant step in capability unlocks exponentially more valuable use cases, justifying the tenfold cost increase. The ultimate prize is tangible: an AI capable of performing the work of a senior engineer could address the $2 trillion global market for software developer wages.
This financial reality forced OpenAI to make a critical choice: build a larger, more intelligent GPT-5 that would be too slow and expensive, or create a same-sized but smarter model. They chose the latter to manage costs and user experience. This decision points to a fundamental shift in AI development. Patel argues that pre-training on internet text is in its "late innings." The future of AI progress and the majority of future compute spending will be in post-training via Reinforcement Learning (RL), creating simulated environments to teach models complex tasks.
According to Patel, the traditional SaaS model is being permanently broken by AI due to a fatal combination of factors:
- High, variable Cost of Goods Sold (COGS): Unpredictable inference costs are destroying traditional software margins.
- Low competitive moats: AI significantly lowers the cost for competitors to replicate product features easily and cheaply.
This pairing prevents companies from achieving the "escape velocity" of profitability that defined the previous software era. The "collapse" may not be a crash in usage, but a painful reckoning for software valuations and business models that are no longer suitable.
Beyond the economic challenges, the AI industry faces massive physical and geopolitical constraints:
- The Electricity and Labor Wall: OpenAI’s planned data center buildout alone will require the equivalent of 23 nuclear power plants. The wages for electricians on data center projects have doubled, highlighting labor bottlenecks.
- The Geopolitical Single Point of Failure: The entire U.S. AI ecosystem is critically dependent on Taiwan for advanced semiconductors. Patel warns that a Chinese blockade would cause the tech economy to "free fall," as everything from AI data centers to cars and refrigerators relies on these chips.
So far, the narrative of mass unemployment due to AI has not materialized. The only clear, data-backed evidence shows sharp declines for freelance graphic designers and copywriters since the advent of ChatGPT. Watch his interview on Invest like the Best for his full breakdown and takes:
Top Takeaways:
- (1:00) - Insight: The "infinite money glitch" (OpenAI paying Oracle, Oracle paying NVIDIA, NVIDIA paying OpenAI) is a meme; the reality is about securing compute.
- (1:20) - Insight: The core problem is OpenAI's "insatiable demand for compute," which must be acquired before revenue-generating business cases can be built.
- (1:44) - Insight: The AI race is a "game of the richest people in the world" (Zuckerberg, Google, Elon). OpenAI, despite its user base, is at risk of being "too small to matter" because it lacks compute.
- (2:13) - Prediction: If OpenAI doesn't secure a position "among the most compute," they will be beaten.
- (2:19) - Insight: OpenAI's original "magic" was simply its foresight and willingness to spend vastly more compute on a single model run (GPT-3/4) than anyone else.
- (2:43) - Example: Mark Zuckerberg's $30 billion deal with Apollo for a single data center illustrates the immense capital required, and that's just for the physical building, not the chips inside.
- (4:01) - Insight: OpenAI needs allies like Oracle to spend the massive capex ahead of the curve, trusting that OpenAI will be able to pay the "rental income" later.
- (4:40) - Statistic: One gigawatt of data center capacity costs $10-15 billion per year on a five-year deal, a $50-75 billion commitment.
- (5:02) - Insight: Sam Altman is asking for 10+ gigawatts, creating a fundamental problem of "who is the balance sheet for this."
- (5:26) - Analysis: Oracle is making a "massive bet" on OpenAI, citing a $300 billion deal. This is a huge gamble against OpenAI's ~$16B current ARR. If it works, Oracle profits $100B; if not, they face huge debt.
- (6:08) - Insight: NVIDIA has a strategic problem: Google and Amazon will fund capex for their own chips (TPUs, Trainium) but not necessarily for Nvidia's GPUs.
- (13:08) - Analysis: The NVIDIA-OpenAI deal mechanics: For a $50B data center, NVIDIA invests $10B in equity. NVIDIA receives ~$35B of that $50B as a supplier.
- (13:51) - Takeaway: This deal allows OpenAI to pay for compute with equity, lets NVIDIA "lower their prices without lowering their prices," and secures NVIDIA's capex dollars upfront by investing in a key customer.
- (3:38) - Insight (Scaling Laws): AI improvement is on a log-log scale (10x compute for the next tier). This looks like diminishing returns, but the value jump is like the difference "a six-year-old versus a 16-year-old."
- (7:33) - Point of View: The speaker confidently states it is not a diminishing return curve, and scaling laws will continue to hold.
- (8:03) - Insight: Bigger models aren't always released. GPT-4.5 was smarter but couldn't be served at a reasonable cost or speed.
- (8:20) - Example: Anthropic's revenue mostly comes from the faster "Sonnet" model, not the smarter "Opus" model, because "no one wants to use a slow model." User experience is key.
- (8:48) - Risk: The financial community is terrified because the moment scaling stops, the hundreds of billions spent on compute will have zero ROI.
- (10:44) - Statistic: Anthropic's revenue ramp (under $1B to $7-8B) is the fastest ever, and it's "basically all code related."
- (14:35) - Insight: This is "about the highest stakes like capitalism game of all time."
- (14:48) - Statistic: Token demand is "doubling every two months or something crazy."
- (15:11) - Tangent: The speaker is trying to rebrand the "economics of tokens" as "Tokconomics" to replace the crypto term.
- (15:47) - Insight (The Serving Trilemma): 1 gigawatt of capacity can serve 1000x of a bad model, 1x of a good model, or 0.1x of an amazing model.
- (17:11) - Insight: It takes time for users to adopt new AI capabilities. GPT-3 was ignored, but ChatGPT (using GPT-3.5) captured attention.
- (19:06) - Insight: OpenAI's strategy (GPT-4 to 4-Turbo to 4o) was to make the model smaller and cheaper at a similar quality, precisely to serve more users and move them up the adoption curve.
- (19:50) - Insight: GPT-5 was not made "way bigger" because they wouldn't be able to serve anyone, and the slowness would kill adoption. It's roughly the same size as 4o.
- (20:13) - Problem: APIs and services are "rate limited" because companies cannot serve all the existing demand, which stalls the adoption curve.
- (21:30) - Problem: Token demand is doubling every two months, but hardware capacity isn't. This is only sustainable because algorithmic improvements are tanking the cost for a given level of intelligence.
- (22:53) - Point of View: If given a "magic button" to fix one bottleneck, capacity/cost is more important than latency.
- (23:45) - Story: The speaker personally uses the "dumber" Anthropic Sonnet model over the smarter Opus model simply because Opus is too slow, and "my time's worth something."
- (24:10) - Concept: "Over-parameterization" means models, like humans, will memorize data before they generalize or "understand" it.
- (24:42) - Concept: Models experience "grokking," an "aha moment" where they shift from memorization to genuine understanding.
- (25:06) - Insight: The main challenge today is not making models bigger; it's generating new data for tasks that don't exist on the internet (e.g., advanced spreadsheet macros).
- (25:45) - Insight: This data gap is being filled by Reinforcement Learning (RL) in "environments"—simulated worlds where the AI can learn.
- (26:25) - Example: Startups are building "fake Amazon" environments to teach models how to click and purchase items, or environments for cleaning dirty data.
- (27:27) - Example: Models "hill climbed" on math puzzles, learning to use Python to solve problems, not just know answers.
- (30:35) - Forecast: In terms of progress, we are in the "late innings on text" for pre-training data but have "like thrown the first ball" on RL environments.
- (30:48) - Analogy: A baby sticking its hand in its mouth is calibrating its senses. Models are "so so early" in this kind of real-world learning.
- (31:40) - Point of View: Some, like Elon Musk, believe "embodiment" (placing AI in a robot) is required for AGI, as it needs physical interaction to understand concepts.
- (32:27) - Forecast: The average person will feel the biggest change when models shift from organizing information to doing things (e.g., "order me this vitamin and it's just done").
- (33:28) - Statistic: Over 10% of Etsy's traffic now comes "straight from GPT."
- (33:40) - Forecast: The clearest monetization path is models making purchases for you and taking a small "take rate," becoming the new Visa.
- (35:50) - Insight: "Reasoning" is just a way to spend more compute (brain cycles) on a task without making the model itself bigger.
- (39:50) - Insight: Models "suck at" infinite context because human memory is "sparse"—we collapse information into dense, fundamental meaning. Models haven't achieved this yet.
- (41:24) - Tangent: Human memory is "fake"; we remember the picture we invented of the memory, which morphs over time, not the actual event.
- (42:25) - Point of View: Models don't need to work like humans. They can use an external "database that it writes stuff in," just as humans use a calendar or notes.
- (43:17) - Example: OpenAI's "Deep Research" feature works this way. It runs for 45 minutes, "writes something down elsewhere" (compressing information), and then refers back to its compressed notes to write the final memo.
- (44:02) - Insight: The need for "millions of GPUs" is not for one giant model, but to "try a bajillion different things" in research because no one knows what will work.
- (46:47) - Insight: We don't need "digital god" AI. The value from simply automating existing software development (like COBOL mainframe migration) would be a "godsend" for the economy.
- (48:01) - Insight: Today's robotics (e.g., picking up a cup) is "impossible for a model" due to the extreme dexterity and tactile feedback required.
- (48:19) - Analogy: A human swishing a wine glass is an incredibly complex feedback loop that models are "nowhere close" to.
- (48:59) - Forecast: For robotics, "we haven't even left the dugout."
- (49:19) - Forecast: AI may soon automate parts of the ML research function, squeezing the number of essential humans down to a tiny few.
- (49:43) - Insight: A billion-dollar researcher salary makes sense. If they are running experiments on a $100B compute cluster, a 5% efficiency gain saves billions in compute time and inference costs.
- (50:53) - Insight: Adding more people to ML research actually slows it down; it's driven by a few individuals with "gut feel."
- (51:30) - Story: A friend at OpenAI gets "viscerally angry" thinking about how many H100s Meta is wasting.
- (52:05) - Takeaway (Run's Idea): The talent war should be national. The US should "acqui-hire" all the best people with process knowledge (e.g., in Shenzhen).
- (54:35) - Story (Jensen Huang): "The reason America is rich... is because we've exported all the labor, but we've kept all the value."
- (56:13) - Analogy: ML research is the exact same as semiconductor manufacturing. Both involve tuning "a thousand different knobs" in an impossibly large search space, relying on intuition and expensive R&D (wasting compute/wafers) to find the next node.
- (58:45) - Insight: In the Cursor/Anthropic relationship, Cursor (app) pays Anthropic (model), who pays NVIDIA (hardware). But Cursor gets all the user data and can switch providers, giving it power. Everyone is "frenemies."
- (1:01:42) - Analysis: In 2024, Microsoft "backed down" from its 2023 "own the world" stance, likely realizing OpenAI couldn't pay for $300B in compute, and let them go to Oracle.
- (1:02:42) - Insight: The Microsoft/OpenAI deal is complicated by an "AGI clause" where Microsoft loses IP rights, but the definition of AGI "always moves."
- (1:04:22) - Insight: NVIDIA can't make major acquisitions, and stock buybacks are for "losers" (admitting you have no better use for capital).
- (1:05:02) - Takeaway: NVIDIA is now "using its balance sheet to win" by offering demand guarantees and backstopping compute clusters to lock in GPU demand.
- (1:05:45) - Insight: NVIDIA "love[s] when venture capitalists fund a company and then 70% of their round is spent on compute."
- (1:09:03) - Prediction: If models stop improving, the massive compute overbuilding will cause a recession in the US, Taiwan, and Korea.
- (1:10:10) - Counterpoint: This isn't a bubble like Tulips because it's funded by the "strongest balance sheets in the world" (Meta, Google), who can "pull the plug at any point."
- (1:10:19) - Example: Microsoft did pull the plug, saw demand was real, and had to "plug it back in" by signing a $19B deal with a new provider (Nebius).
- (1:12:54) - Insight: The "neocloud" (GPU rental) business is "amazing or terrible." It's terrible if you sell short-term contracts (you'll be obsolete) but amazing if you sign long-term deals with companies like Microsoft.
- (1:17:19) - Takeaway: In the AI value chain, "NVIDIA's holding no risk. Everyone in the middle's got a lot of risk."
- (1:19:34) - Story: The speaker's own business (tracking data centers via satellite imagery) "would not have been possible" without AI, turning a hypothetical 100-person job into a 3-person one.
- (1:41:07) - Startup (Periodic Labs): An interesting startup applying the RL paradigm to the physical world (material science, battery chemistry), creating a feedback loop between simulation and real-world lab tests.
- (1:43:33) - Point of View: The speaker is not bullish on AI accelerator startups (NVIDIA competitors); it's "too hard" and "too capex intensive."
- (1:44:01) - Insight: The real hardware innovation is in the "old" parts of the supply chain, like power transformers, which haven't changed in 100 years.
- (1:55:52) - Insight: The traditional SaaS business model (low COGS, high acquisition cost) is breaking.
- (1:56:47) - Analogy: China never developed a big SaaS market because it was cheaper to build software than to buy/rent it. AI is making this true for everyone.
- (1:58:23) - Prediction (Death of SaaS): AI adds a "humongous COGS" (compute) to software, while simultaneously tanking development costs. This means new software-only businesses will struggle to hit "escape velocity" due to high costs and high competition.
- (1:22:34) - Insight: AI power consumption is "literally nothing" (2% of US grid). The problem is "we haven't built power in 40 years," and the supply chains are broken.
- (1:23:26) - Statistic: A single 2-gigawatt OpenAI data center will consume as much power as the entire city of Philadelphia.
- (1:25:14) - Tangent: One company is building a power plant by putting diesel truck engines in parallel because the industrial capacity for them is huge and untapped.
- (1:26:13) - Statistic: Wages for mobile electricians working on data centers have "doubled"; it's like being a West Texas fracking guy in 2015.
- (1:28:05) - Insight: Grids are creating rules to cut power to data centers, forcing them onto backup generators. This, in turn, may violate air permits if the generators run for too long.
- (1:30:11) - Point of View: If not for the AI boom, the US "probably would be behind China" as the world hegemon by 2030.
- (1:30:36) - Point of View: "The US really really needs AI" to accelerate GDP. "Once you start talking about dividing the pie, you're screwed."
- (1:31:56) - Insight: China doesn't need AI to win; they are playing the long game of dominating industrial supply chains (steel, solar, etc.).
- (1:33:42) - Insight: China's strategy is insular (supply chain security) and they've dumped $400-500B into their semiconductor ecosystem, far more than the US CHIPS Act.
- (1:35:25) - Statistic: Bytedance is the 2nd or 3rd largest user of GPUs in the world.
- (1:38:05) - Risk (Doomsday Scenario): If China blockades Taiwan, the US economy "free falls" because "we can't make refrigerators without Taiwanese chips," let alone AI data centers.
- (1:39:11) - Takeaway: You can't be bearish on TSMC (due to geopolitical risk) without also being bearish on Apple, Amazon, Google, and Microsoft, who depend on it.
Conclusion: A Real Revolution, Financed by a Generational Bubble
After dissecting the arguments from every angle, the evidence points not to a simple bull or bear victory, but to a complex paradox: we are witnessing a genuine technological revolution being financed by what is almost certainly an unsustainable financial bubble. The bull and bear cases are not mutually exclusive; they are two sides of the same coin.
Where the Bulls (Andreessen, Huang, Pompliano, Theo) Are Factually Correct:
- Adoption Speed is a Historical Anomaly: Marc Andreessen’s core observation is undeniable. The latest figures of 800 million users for OpenAI are tangible proof of genuine product-market fit and a level of consumer interest that no previous technology has achieved so quickly. His thesis that the product is already "monumentally amazing" compared to the early, clunky internet is valid.
- The Industrial Vision is Not a Fantasy: Jensen Huang's case for a multi-trillion-dollar "Great Refresh" is confirmed by the public financials of the world's largest companies. Microsoft's cloud gross margins are actively compressing "driven by scaling our AI infrastructure," and Alphabet has committed a staggering $85 billion in 2025 CapEx. As Huang argues, this is "existential spending," not speculative froth.
- The Underlying Cost of Intelligence is Falling: The bull case for long-term profitability is supported by hard data. Martin Alderson’s model of a 1000x cost asymmetry between input and output tokens, combined with independent benchmarks from Epoch AI showing the cost-per-performance is collapsing by 9x to 900x per year, provides a clear, data-backed reason to believe the raw cost of intelligence is on a deflationary curve. This is the "steep, ongoing cost curve" that gives investors hope.
- The Investor Framework is Logically Sound (On Paper): The model articulated by Anthropic's Dario Amodei and illustrated by Theo—which reframes today’s losses as R&D for a future, exponentially better product—is the core of all venture capital. Given OpenAI's >200% YoY revenue growth, the assumption of a revenue multiplier (4x) eventually outpacing a cost multiplier (3x) is, at least, mathematically plausible.
Where the Bears (Zitron, Garran, Szyszka, Ding) Are Factually Correct:
- The Current Profit & Loss is Catastrophic: The numbers are non-negotiable. OpenAI is spending ~$3 for every $1 it earns, running at a potential $20 billion annual loss run rate. As Ed Zitron argues, no business can sustain this without an infinite stream of capital. The revenue/cost gap for the entire industry is enormous, with an estimated $500B+ in infrastructure spending to support just $61B in 2025 revenue.
- The Financial Structure is a Bubble: Zitron's "circular funding" thesis is quantitatively backed by Julien Garran's most damning data point: NVIDIA's receivables have grown 626% in 30 months, a figure that dwarfs the 138% Cisco saw before its dot-com crash. This strongly suggests that a significant portion of "demand" is financial engineering—a wealth transfer mechanism to a single shovel seller.
- Real-World Unit Economics Are Structurally Unprofitable: Garran's data center math provides a devastating real-world rebuttal to Alderson's theoretical model. If it costs $6.31/hour for an operator to break even on a Blackwell GPU but the public market rate is only $3.79/hour, the ecosystem is guaranteed to lose money. This confirms the "Token Short Squeeze" theory from Szyszka and Ding and explains why application companies built on APIs are getting crushed.
- Adoption is Shallow and Lacks Economic Value: The bulls' "800M users" figure is powerfully countered by the bears' data: while AI shows task-level productivity gains, 74-80% of enterprises fail to capture business value at this stage. This is the key disconnect: widespread use, yes, but a widespread failure to monetize or impact the bottom line, also yes. The bill will come due eventually, and when it does, the lights go off and the ice cream melts.
While the bull case for the long-term potential of AI is strong, the bear case on the current market structure is built on a mountain of verifiable, real-world financial data. Therefore, Zitron and Garran's core economic arguments are more compelling in the immediate term. Economics ultimately matters more than adoption. A product that is widely used but cannot be profitably scaled is a failed business. The critical flaw in the bull case is the assumption that profits will automatically follow hype. As the ghosts of Pets.com and MoviePass remind us, they often don't.
The question isn't "will AI be used?" but "can these AI businesses ever be profitable at a scale that justifies the capital invested?" The evidence suggests: not with current technology and at current prices.
Something must give. This leads to a probabilistic conclusion:
- Probability of a Major Correction/Bubble Collapse (Zitron/Garran's Thesis): 70%
- Probability of Sustained, Uninterrupted Growth (Andreessen/Huang's Thesis): 30%
The 30% bull scenario is not impossible, but it requires one of three things to happen:
- Miraculous Cost Reductions: The deflationary curve tracked by Epoch AI must accelerate even faster, collapsing the cost of inference by another 90%+.
- Discovery of a "Killer Use Case": An application must emerge—perhaps addressing the $2 trillion global market for software developers, as Dylan Patel suggests—that is so valuable it justifies 10x higher prices.
- A New Breakthrough Beyond Scaling: A new architectural paradigm must emerge that sidesteps Garran's "scaling wall," delivering the next leap in capability without an exponential increase in cost.
Now, if you want some hints as far as what needs to get resolved from a technology perspective to push us closer towards AGI (AI that will justify these levels of spending), check out the recent interview Dwarkesh Patel did with Andrej Karpathy, which we break down here.
The most likely outcome is a painful synthesis predicted by Dylan Patel: The Death of the SaaS Model, where the application layer faces a brutal washout. Ultimately, we are in a speculative, debt-fueled AI infrastructure bubble, and a painful correction is likely. The core tension is that inference, not training, is the real long-term cost wall; if scaling laws stop improving, the massive misallocation of capital could trigger a US recession.
However, the long-term future likely belongs to the bulls, but only after the bears get their reckoning. A popping bubble doesn't negate the underlying technology. Just as the telecom bubble of the 1990s left behind cheap fiber-optic cable that enabled the Web 2.0 revolution, the enduring legacy of this era will be the historic buildout of computational infrastructure it financed. The companies that survive the coming winter will inherit a world with unimaginable computing resources. The real AI revolution may not be happening now, but the bubble of today is doing the inefficient but necessary work of building the world where it can happen tomorrow.
A bolder, more cynical conclusion (that might offend some industry folks reading this) would go like this: the crux of the current AI bubble market dynamics is driven not by the needs of the underlying technology, but by the existential needs of its two largest players.
The colossal spending across the industry serves to justify NVIDIA's unprecedented market capitalization, creating a feedback loop that allows it to charge near-monopolistic prices for its chips, while insulating it from the market forces that would normally compress its margins. Simultaneously, OpenAI's frantic fundraising is a desperate bid for sovereignty—a race to build its own data centers and escape being completely consumed by its hyperscaler partners like Microsoft that currently control its destiny.
The combined effect of these two forces—NVIDIA's need to defy economic gravity and OpenAI's need to escape it—has placed theentire global economy in the position of a frog in a massive pot of water. The pot sits across two burners. One is controlled by the chipmaker, the other by the model-maker, and both are slowly turning up the heat.
Where Is All This Going? A Massive Inflection Point
This brief video from Peter Leyden is a good overview of where things go from here, because it compellingly frames AI not as a mere tool, but as a "world historic" tipping point, comparable to the steam engine's amplification of our physical power, that is now poised to amplify our mental power and launch the "Age of AI." Even if we're in a bubble now, the reason things feel so messy and complicated is because we are in the direct middle of a major inflection point that will likely define the rest of the decade.
- (1:05) Leyden recounts his time at Wired magazine in the 1990s, drawing a parallel between people's inability to grasp the potential of the internet, email, and "goofball startups like Amazon" back then to how people may be underestimating AI's impact today.
- (8:22) Artificial Intelligence is identified as the most obvious of three "world historic, game-changing technologies" (alongside clean energy and bioengineering) that are beginning to scale and trigger a societal tipping point.
- (8:39) The release of ChatGPT 3.5 in November 2022 will be looked back on as a "world historic moment" and the "starting gun" for what will be understood as the "Age of AI."
- (8:57) The "Age of AI" is framed as a fundamental "step change in our abilities" comparable to humanity entering the Bronze Age or the Iron Age, marking a threshold that, once crossed, "you don't go back" from.
- (9:17) We are about to witness an "explosion" in the "amplification of our mental powers" driven by digital computers and, most recently, AI.
- (9:20) Leyden makes a direct comparison: the amplification of our mental powers by AI will be "very similar" to the amplification of our physical powers by mechanical engines (like steam), which created the prosperity of the modern world.
- (10:55) Leyden notes that the earliest breakthroughs in generative AI (about 15 years prior to the talk) occurred around the same time as the major breakthroughs in bioengineering (CRISPR).
- (12:47) AI is part of the "foundation" for a much larger societal shift—a "once in 80-year reinvention" and potentially the "early days of building a 21st-century civilization."
- (13:45) Leyden predicts that new technologies, implicitly including AI and digital systems, will drive a shift from our current "representative democracy" to a new form of "digital democracy."
- (14:06) He forecasts a move away from the nation-state model toward "some kind of global governance" (we'd argue an evolution into network states are equally likely) that will be necessary to coordinate 10 billion people on a planet managing the power of technologies like AI.
It seems inevitable that at some point, in whatever form it ultimately takes, AI is here to stay. The chaos, the hype, and the fear are simply the sounds of a new world being built. To quote Jacqueline Cane, a brilliant commenter on the video above: "Paleolithic emotions, medeival institutions and god-like technologies...what could go wrong?"







