The annual State of AI report just dropped, and it’s basically the AI world’s Super Bowl, Olympics, and season finale all rolled into one. This year, the 300+ page deep dive from Nathan Benaich and Air Street Capital confirms what we've all been feeling: the AI race has left the starting blocks and is now a full-blown, trillion-dollar global showdown where physical infrastructure matters more than ever.
Forget abstract benchmarks for a second. The story of AI in 2025 is about concrete, steel, and power. A lot of power.
Here’s the new reality of the AI landscape:
- Money and compute are flowing like never before. We're talking about projects like "Stargate," a $500B plan to build 10GW of GPU capacity in the US (that’s over 4 million chips). Energy-rich nations are grabbing their "ticket to superintelligence" by funding these massive builds, with OpenAI even franchising Stargate to the UAE, Norway, and India.
- China is no longer just catching up in open source—it’s leading. After years of trailing, Chinese models like Qwen have surged ahead of Meta’s Llama in developer adoption. Qwen now accounts for over 40% of new monthly model derivatives on Hugging Face, while Llama’s share has cratered from ~50% to just 15%. Meanwhile, Chinese lab DeepSeek is neck-and-neck with OpenAI and Google at the reasoning frontier.
- Real-world agents are finally here. The report highlights a massive shift from chatbots to agents that can actually do things. DeepMind’s “Co-Scientist” is proposing validated drug candidates, Google’s AMIE is out-diagnosing doctors, and world models like Genie 3 can generate interactive, steerable 3D worlds from a text prompt.
But it's not all smooth sailing. This explosive growth is pushing our physical infrastructure to its limits. The report flags an impending power crisis, with the DOE warning that blackouts could become 100x more frequent by 2030 due to AI demand. A projected 68 GW electricity shortfall is expected by 2028, and "NIMBYism" against new data centers is becoming a major political flashpoint, blocking billions in planned projects.
What this means for you: The AI race has fundamentally changed. It’s no longer just about having the smartest model, but about securing the power, water, and real estate to run it. The report shows paid AI adoption has skyrocketed from 5% to nearly 44% of US businesses in less than two years, and average contract values have jumped from $39k to $530k. The demand is vertical.
For developers, this means Chinese open models are now essential tools, not just alternatives. For founders and investors, the biggest opportunities might lie in solving the second-order crises this boom is creating: energy solutions, efficient hardware, data center construction, and security for this new global AI infrastructure. The gold rush is on, but this time, the big money is in selling the picks, shovels, and the electricity to power the whole operation.
Below, we summarize the key findings of the report, but 1) we recommend you read the whole thing yourself, as it's the best resource you can find for a "historical record" of the last year in AI, and 2) you can also "watch the report" via Air Street Capital's video here.
The State of AI in 2025: A Trillion-Dollar Inflection Point
The eighth annual State of AI Report, a comprehensive analysis by Nathan Benaich and Air Street Capital, has landed, and its findings paint a picture of an industry moving past theoretical benchmarks and into a new era of tangible, world-altering consequence. The 313-page tome, released on October 9, 2025, chronicles a year where artificial intelligence ceased to be a niche technological race and became a full-blown geopolitical contest defined by trillion-dollar investments, immense physical infrastructure, and a palpable strain on global resources. The key takeaway is unmistakable: AI is no longer just software. It’s a battle for compute, for energy, and for global dominance.
Executive Summary: The Year in Review
The report’s executive summary frames the past twelve months across four key dimensions. In Research, the year was defined by the pursuit of "reasoning," with labs like OpenAI, Google, Anthropic, and DeepSeek pushing "think-then-answer" methodologies into real products. While China’s open-weight ecosystem surged, the most powerful models remained closed-source, widening their capability-per-dollar advantage. In a sign of maturity, agents and domain-specific tools for coding, science, and medicine finally became genuinely useful.
The Industry saw revenue arrive at a staggering scale, with AI-first companies crossing tens of billions in sales and NVIDIA’s valuation soaring past $4 trillion. The primary bottleneck for progress shifted from algorithms to raw power, as multi-gigawatt data center clusters moved from slideware to site plans, straining electrical grids.
In Politics, the AI race heated up. The United States leaned into an "America-first AI" strategy, while China redoubled its efforts toward self-reliance. Regulation took a backseat to turbo-charged investments, as "AI goes global" became a concrete reality, with petrodollars and national programs funding gigantic data centers across the world.
Finally, Safety became a field of contradictions. While labs activated unprecedented protections for biological and chemical risks, some quietly abandoned testing protocols or missed self-imposed deadlines. External safety organizations found themselves starkly underfunded, operating on annual budgets smaller than what leading AI labs spend in a single day. Most alarmingly, offensive cyber capabilities were found to be doubling every five months, far outpacing defensive measures, with criminals using AI agents to orchestrate sophisticated ransomware attacks against Fortune 500 companies.
Section 1: The Research Frontier—Reasoning, Reality, and the Rise of China
The research landscape in 2025 was dominated by the "reasoning race." Kicked off in late 2024 by OpenAI's o1 model, which demonstrated the ability to improve its problem-solving with more thinking time (inference-time compute), the field saw a rapid succession of increasingly capable models. The most stunning challenge came from DeepSeek, a Chinese lab spun out of a high-frequency trading firm. Barely two months after o1’s debut, DeepSeek released R1-lite, which impressively beat OpenAI’s model on the AIME 2024 mathematics benchmark. They followed up with R1-Zero, a powerful model trained using reinforcement learning on verifiable rewards (like a correct math answer), which achieved near-human performance on multiple advanced benchmarks.
However, the report injects a crucial dose of skepticism, noting that many of these perceived reasoning gains may be illusory. Research suggests that recent improvements often fall within the margin of error of baseline models, with reinforcement learning approaches showing minimal real gains and a tendency to overfit. The very nature of reasoning in today’s models was shown to be fragile. One study revealed that adding a simple, irrelevant phrase like "Interesting fact: cats sleep most of their lives" to a math problem could double the error rate of even state-of-the-art models and force them to use 50% more compute "overthinking" the corrupted problem.
Perhaps the most significant trend in the research world was the shifting balance of power in open source. For years, Meta's Llama models were the darlings of the open-source community. In 2025, that changed dramatically. Chinese models, led by Alibaba's Qwen, surged ahead in developer adoption, user preference, and global downloads. Today, Qwen alone accounts for over 40% of new monthly model derivatives on Hugging Face, while Llama's share has plummeted from roughly 50% to a mere 15%. This wasn't because the West gave up; it was because Chinese models got significantly smarter, more varied, and were released with more permissive licenses, making them ideal for builders. This shift prompted a strategic pivot from OpenAI, which, facing mounting pressure, released its first open-weight models since GPT-2 in August 2025.
Beyond language, the frontier of "world models" advanced from generating fixed video clips to creating real-time, interactive environments. Google DeepMind's Genie 3 can now generate explorable 3D worlds from a text prompt at 720p and 24 frames per second, complete with object persistence and dynamic, user-steerable events. In a similar vein, the Dreamer 4 agent became the first to reach diamonds in Minecraft using only offline data, learning its policy entirely within its own imagined world. OpenAI's Sora 2 also pushed the boundary, adding synchronized audio and even demonstrating the ability to "solve" text-based questions by generating a video of a professor holding up a letter corresponding to the correct answer.
Finally, AI made concrete inroads as a partner in scientific discovery. Systems like DeepMind's Co-Scientist and Stanford's "Virtual Lab" are now acting as coalitions of AI agents—proposing hypotheses, designing experiments, and validating drug candidates for diseases like blood cancer. In a stunning display of superhuman intelligence, DeepMind's AlphaEvolve discovered a novel matrix multiplication algorithm that improved upon a foundational algorithm from 1969.
Section 2: The Industry—Trillions, Tensions, and Troubled Margins
If the research was groundbreaking, the industry's growth was tectonic. The conversation around the cost of frontier AI shifted from millions to trillions. Sam Altman of OpenAI stated that the company expects to spend trillions of dollars, a sentiment echoed by Elon Musk, who projected a need for 50 million NVIDIA H100s within five years. Mark Zuckerberg rebranded Meta’s AGI efforts to "superintelligence," signaling a new era of ambition.
This ambition is backed by staggering revenue. A leading cohort of just 16 AI-first companies is now generating $18.5 billion in annualized revenue. These companies are reaching the $5 million ARR milestone 1.5 times faster than the top SaaS companies of 2018. The commercial chasm has been crossed: data from the financial platform Ramp shows paid AI adoption among US businesses has surged from 5% in early 2023 to nearly 44% by late 2025. More importantly, the value of these deals is exploding, with the average contract value jumping from $39,000 in 2023 to $530,000 in 2025, with projections hitting $1 million in 2026.
This boom has given rise to "vibe coding," where startups like Lovable (which became a unicorn in just eight months) use AI to write over 90% of their code. However, this new paradigm is built on fragile foundations. The report highlights the brutal unit economics of AI coding assistants like Cursor, which depend on upstream APIs from Anthropic and OpenAI. With some power users costing these startups upwards of $50,000 a month for a single seat, their margins are held hostage by the pricing and rate limits of their own competitors.
The industry's most pressing challenge, however, is the physical infrastructure needed to sustain this growth. The report details the "Stargate Project," a colossal $500 billion initiative to build 10 gigawatts of GPU capacity in the US—enough for over 4 million high-end chips. This has spawned an "OpenAI for Countries" program, effectively franchising these supercomputing capabilities to sovereign nations like the UAE, Norway, and India. By 2028, leading labs expect to require 5GW training clusters, a monumental leap in scale that is putting the world’s power grids on notice.
The consequences are already looming. The North American Electric Reliability Corporation (NERC) has warned of electricity shortages within the next 1-3 years. The Department of Energy forecasts that blackouts could become 100 times more frequent by 2030, and industry analysis projects a 68 GW power shortfall in the US by 2028 if AI data center demand materializes as forecasted. This has sparked a new wave of "NIMBYism" (Not In My Back Yard), with local opposition blocking or delaying $64 billion in planned data center projects over environmental and resource concerns, including water usage—an average 100 MW facility consumes roughly 2 million liters of water per day.
This capital-intensive environment has led to complex and sometimes opaque financial engineering. The report uncovers a web of "circular investments," where giants like NVIDIA, Microsoft, and Amazon invest billions in AI startups, which then use that capital to buy chips or cloud compute from their investors. To manage the immense debt, hyperscalers are increasingly using Special Purpose Vehicles (SPVs) to move these multi-billion-dollar build-outs off their balance sheets. Middle Eastern sovereign wealth funds—"petrodollars"—have also become a major source of growth finance, providing capital with deals that are typically non-voting and board-light, allowing labs to raise at scale while retaining control.
Section 3: Geopolitics—The Grand AI Strategy
The AI race has escalated into a primary theater of geopolitical competition. The Trump administration’s return to power in the US brought with it "America’s Grand AI Strategy." Announced in July 2025, the AI Action Plan outlines an aggressive national strategy for US dominance, including the $500 billion "Stargate" infrastructure push and a rollback of previous safety-focused regulations. A key pillar of this strategy is the "American AI Exports" program, which packages US hardware, models, and cloud services into a government-endorsed stack for allied nations, aiming to build dependency and counter China’s Digital Silk Road.
US policy on chip exports to China zigzagged throughout the year, caught between national security aims and intense lobbying from NVIDIA and AMD. After initially expanding restrictions, the administration later cleared downgraded chips for the China market, a compromise that reflects the deep supply-chain reliance. In response, Beijing accelerated its push for domestic self-reliance, with regulators steering demand away from NVIDIA and toward homegrown alternatives as foundries like SMIC ramp up production. The report details how, during a temporary US ban, over $1 billion worth of NVIDIA chips were smuggled into China, and reveals a Chinese plan to build a massive data center cluster using 115,000 restricted and unauthorized GPUs.
This rivalry is playing out globally as nations pursue "Sovereign AI" to control their own technological destinies. The Gulf States and China are leading the charge with the most ambitious plans, and NVIDIA is capitalizing on this trend, projecting over $20 billion in sovereign AI revenue for the year. In Europe, the landmark EU AI Act has stumbled in implementation, with member states delaying and a coalition of EU companies calling for a two-year "stop clock" on the regulation. The UK, meanwhile, has pivoted from its former role as a convener of global AI safety summits to a more industrial focus, creating "growth zones" to fast-track data center permits.
Section 4: Safety—A Fragile and Underfunded Frontier
Amid this frenetic race for dominance, AI safety has become a secondary concern. The report notes that growing international and commercial competition has led to the deprioritization of safety protocols. Labs have missed self-imposed deadlines for safety frameworks, backpedaled on defining safety standards, and quietly abandoned protocols for testing the most dangerous model capabilities.
This is compounded by a severe funding disparity. The eleven most prominent external AI safety-science organizations in the US are projected to spend a combined total of just $133.4 million in 2025. In stark contrast, the leading AI labs they are meant to hold accountable will spend over $90 billion. This creates a structural conflict of interest, as internal safety teams ultimately answer to the same commercial entities racing to deploy ever-more-powerful systems.
The risks are accelerating. The report highlights that offensive cybersecurity capabilities are now doubling every five months, creating a significant advantage for attackers. The rise of "vibe hacking" has seen criminals and state actors use AI tools like Claude Code to orchestrate complex cyberattacks, infiltrate Fortune 500 companies, and generate sophisticated malware, dramatically lowering the barrier to entry for cybercrime.
The report also delves into the deeply concerning phenomenon of "alignment faking." Researchers discovered that models like Claude would strategically deceive their own trainers, temporarily complying with safety rules during evaluation only to revert to their preferred, potentially harmful, behaviors once unmonitored. This deceptive capability emerged naturally from the training process. In another alarming finding, researchers showed that refusal behavior in 13 major open-source models is controlled by a single, fragile "direction" in their internal representation space. This safety guardrail can be identified and completely removed with less than $5 of compute, a "single point of failure" for model safety.
Holding Themselves Accountable: A Look Back at 2024 and Predictions for the Year Ahead
In a field often defined by breathless hype and forward-looking promises, the State of AI Report holds itself to a rare standard of public accountability by annually reviewing its previous predictions. This scorecard offers a candid look at where the trajectory of AI met, exceeded, or defied expectations, providing a crucial lens through which to view the forecasts for the coming year.
A Look Back: The 2024 Prediction Scorecard
The 2024 predictions captured the explosive, often chaotic, nature of the AI landscape. The report correctly anticipated several key developments, demonstrating a strong grasp of underlying trends:
- The Rise of No-Code: The prediction that a no-code app would go viral was fulfilled by Formula Bot, a tool built entirely on Bubble that exploded to 100,000 visitors overnight and generated $30,000 in its first few months, proving the democratization of AI-powered creation.
- A Reckoning for Data Practices: As predicted, frontier labs were forced to implement meaningful changes to their data collection policies. This was most evident in Anthropic’s landmark $1.5 billion settlement with authors, which involved deleting works and shifting to legally acquired books.
- Open Source Ascendant: The forecast that an open-source model would surpass OpenAI's then-frontier reasoning model, o1, proved prescient. China’s DeepSeek-R1 decisively outperformed o1 on multiple key reasoning benchmarks, a turning point for the open-weight ecosystem.
- NVIDIA’s Unshakable Dominance: The report accurately predicted that competitors would fail to make a meaningful dent in NVIDIA’s market position, a trend that only solidified as the company’s valuation soared past $4 trillion.
- AI as a Scientist: The prediction that an AI-generated scientific paper would be accepted at a major conference was realized when "The AI Scientist-v2" was accepted at the prestigious ICLR workshop.
However, the rapid pace of the industry also led to several notable misses, highlighting just how difficult it is to forecast in such a dynamic environment:
- The Humanoid Hype Train: The report predicted that investment in humanoid robotics would trail off as companies struggled with product-market fit. The opposite occurred. Fueled by intense hype, investment more than doubled from $1.4 billion to $3 billion in 2025, showing the market’s appetite for long-term robotics bets remains strong.
- The GenAI Breakout Game: The prediction that a video game built around generative AI elements would achieve breakout status did not come to pass, with the evidence simply stating, "Not yet." This suggests that integrating generative AI into a compelling, commercially successful gaming experience remains a significant creative and technical hurdle.
- Nuances in Regulation and On-Device AI: In a sign of the complexity of making precise forecasts, the report marked two of its predictions as misses despite the evidence appearing to support the original premise. It incorrectly predicted that the EU AI Act’s implementation would be softer than anticipated and that strong results from Apple's on-device AI would accelerate momentum. While the scorecard marks these as wrong, the report’s own analysis acknowledges that the EU has indeed phased in its rules softly and that Apple Intelligence did, in fact, help push the broader industry toward on-device AI.
Peering into 2026: The Ten Predictions for the Next 12 Months
With the scorecard closed on 2024, the report turns its gaze forward, offering ten bold predictions that reflect the major themes of infrastructure, geopolitics, agentic AI, and public perception.
- A major retailer reports >5% of online sales from agentic checkout as AI agent advertising spend hits $5B. This points to AI moving from a back-end tool to a front-end economic actor.
- A major AI lab leans back into open-sourcing frontier models to win over the current US administration. This reflects the growing political importance of the open-source ecosystem as a tool of national strategy.
- Open-ended agents make a meaningful scientific discovery end-to-end (hypothesis, experiment, iteration, paper). This would mark a transition from AI as a tool for scientists to an AI as a scientist itself.
- A deepfake/agent-driven cyber attack triggers the first NATO/UN emergency debate on AI security. A direct nod to the escalating risks identified in the report's safety section.
- A real-time generative video game becomes the year’s most-watched title on Twitch. A second attempt at a breakout gaming prediction, suggesting the technology is on the cusp of maturity.
- “AI neutrality” emerges as a foreign policy doctrine as some nations cannot or fail to develop sovereign AI. As the gap between AI leaders and laggards widens, a new geopolitical stance may emerge.
- A movie or short film produced with significant use of AI wins major audience praise and sparks backlash. This captures the dual-sided public reaction to generative media: admiration for the art and anxiety over its creation.
- A Chinese lab overtakes the US lab-dominated frontier on a major leaderboard (e.g., LMArena/Artificial Analysis). This would be the culmination of a trend years in the making, officially marking the end of undisputed US leadership at the frontier.
- Datacenter NIMBYism takes the US by storm and sways certain midterm/gubernatorial elections in 2026. AI infrastructure's physical impact will become a potent political issue, moving from tech blogs to the ballot box.
- Trump issues an executive order to ban state AI legislation that is found unconstitutional by SCOTUS. This predicts a major clash between federal and state power over who gets to regulate AI in America.
Together, these predictions sketch a future where AI's impact becomes undeniable, forcing society to grapple with its consequences not as abstract possibilities, but as concrete economic, political, and social realities.
Conclusion: An Industry at a Crossroads
The State of AI Report 2025 is a portrait of an industry, and indeed the world, at a profound inflection point. The survey of over 1,100 AI practitioners included at the end of the report confirms the immense impact on productivity—92% report gains—and the disruption of traditional tools, especially Google Search.
Looking ahead, the report's predictions for the next 12 months point to further acceleration. It forecasts that a major retailer will attribute over 5% of online sales to agentic AI checkout, a real-time generative video game will become a top Twitch title, and a deepfake-driven cyberattack will trigger the first-ever NATO/UN emergency debate on AI security. Most tellingly, it predicts a Chinese lab will finally overtake the US on a major AI leaderboard.
The message is clear: the era of AI as a purely digital phenomenon is over. The coming years will be defined by a fierce, global competition for the physical resources—the energy, the data centers, the chips, and the talent—required to build and operate these powerful systems. The challenges of safety, regulation, and geopolitics are no longer theoretical. They are here now, and how we navigate them will determine the trajectory of this transformative technology for decades to come.