Research & Development World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • 2025 R&D 100 Award Winners
    • 2025 Professional Award Winners
    • 2025 Special Recognition Winners
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
  • Resources
    • Research Reports
    • Digital Issues
    • Educational Assets
    • Subscribe
    • Video
    • Webinars
    • Content submission guidelines for R&D World
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE

AI’s great compression: 20 charts show vanishing gaps but still-soaring costs

By Brian Buntz | April 12, 2025

This year’s AI Index Report from Stanford University’s Institute for Human-Centered AI (HAI) paints a picture of an AI landscape accelerating at a startling pace, challenging simple narratives of an ‘AI bubble’ or progress deceleration. While the compute used in the largest training runs continues its exponential climb (historically doubling roughly every 3–4 months), the cost to deploy AI is simultaneously plummeting. Inference costs are falling like a rock. Specifically, they dropped over 280-fold in just 18 months for comparable performance. Mirroring this cost-performance dynamic are widening performance gaps between proprietary and open models. Similarly, U.S. and Chinese leaders are locked in intense competition at the frontier. For professionals in R&D, navigating this era of intense acceleration and capability compression requires close attention as the advances continue to have potentially significant repercussions throughout the STEM landscape. Here are standout findings of the report:

1. Hardware performance doubles every 1.9 years

Chart showing annual growth rate of ML hardware performance
A primary engine of AI progress continues to fire on all cylinders. The report finds: “‘AI hardware gets faster, cheaper, and more energy efficient.” It adds: “New research suggests that machine learning hardware performance, measured in 16-bit floating-point operations, has grown 43% annually, doubling every 1.9 years.’ (Page 29, Highlight 8).” This improvement pace enables the training of larger models though there is a simultaneous trend towards extracting more performance through distillation and test-time compute approaches. The trend, however, doesn’t mean that AI clusters’ performance should be compared head-to-head against supercomputers, as the latter uses considerably higher accuracy.

2. South Korea and Luxembourg lead AI patenting per capita

Chart showing AI patents per capita by country
While China’s sheer volume of AI patents is unmatched (69.7% of grants in 2023), focusing on per capita figures reveals unexpected IP hubs. The report notes that “‘South Korea and Luxembourg stand out as top AI patent producers on a per capita basis.’ (Page 29, Highlight 7).” These nations foster environments conducive to generating novel AI-related intellectual property relative to their population size.

3. US produced the most notable AI models in 2024, but China overtook Europe

Chart showing total notable AI models by geographic area from 2003-2024
Historically, the U.S. has been the primary source of breakthrough AI models. This continued in 2024: “‘In 2024, U.S.-based institutions produced 40 notable AI models, significantly surpassing China’s 15 and Europe’s combined total of three.’ (Page 28, Highlight 4).” Yet the data also shows a shift, with China producing five times as many notable models as Europe in 2024.

4. East Asia leads publication volume, US leads in influential research

Chart showing AI publications by world region over time
Looking at research output, a regional divergence appears. “‘In 2023, East Asia and the Pacific led AI research output, accounting for 34.5% of all AI publications, followed by Europe and Central Asia (18.2%) and North America (10.3%).’ (Page 33).” The report, however, also indicates the U.S. remains the leading producer of the most highly cited (top 100) publications (Page 28, Highlight 1), suggesting a potential contrast between research quantity (led by East Asia) and immediate high-impact influence (led by the US).

5. Industry deploys vastly more compute, widening the resource gap

Chart showing training compute distribution by sector
Developing frontier AI models requires immense resources, increasingly concentrating this capability within industry. As the report states, “‘Creating cutting-edge AI models now demands a colossal amount of data, computing power, and financial resources that are not available to academia. Most leading AI models are coming from industry…’ (Page 58).” Figure 1.3.13 (Page 55) shows that industry almost exclusively deploys the highest training compute levels.

6. API access is the dominant release strategy for notable models

Chart showing distribution of notable AI models by access type in 2024
How are cutting-edge models made available? The data shows a clear preference for controlled access. “‘In 2024, API access was the most common release type, with 20 of 61 models made available this way, followed by open weights with restricted use and unreleased models.’ (Page 51).” This strategy allows developers to monetize and control use, but limits deeper inspection and customization.

7. China solidifies lead in AI publication volume

Chart showing AI publication share by country/region over time
Zooming in on national research output quantity, the report confirms China’s established lead: “‘In 2023, China was the global leader in AI article publications, accounting for 23.2% of the total, compared to 15.2% from Europe and 9.2% from [the United States].’ (Page 35).” Since surpassing other regions around 2016-2017, China has maintained its top position in the sheer volume of AI papers published, though as noted earlier, the U.S. currently leads in producing the most-cited research.

8. Performance gap between US and Chinese models nearly vanishes

Chart showing performance gap between top US and China models on benchmarks over time
Perhaps one of the most significant competitive dynamics revealed is the rapid convergence at the performance frontier: “‘The gap closes between Chinese and U.S. models… At the end of 2023, performance gaps on benchmarks such as MMLU… and HumanEval were 17.5… and 31.6 percentage points… By the end of 2024, these margins had narrowed substantially to 0.3… and 3.7 percentage points.’ (Page 15, Highlight 3).” This near-parity signals intensified global competition and suggests Chinese models are rapidly achieving capabilities comparable to their top US counterparts.

9. ImageNet accuracy reaches 91.5%, but progress slows

Chart showing performance progress on the ImageNet computer vision benchmark
AI systems continue to master core visual tasks. Performance on the foundational ImageNet benchmark hit 91.5% top-1 accuracy in late 2024 (Figure 2.2.7). The report, however, cautions that the rate of improvement on such established benchmarks has decelerated compared to previous years. That finding potentially indicates that these specific tests are nearing saturation and may no longer effectively differentiate frontier model capabilities.

10. Object detection (COCO benchmark) scores continue steady climb

Chart showing performance progress on the COCO object detection benchmark
AI’s ability to precisely identify and locate objects within complex scenes keeps advancing. Performance on the COCO benchmark reached a mean Average Precision (mAP) of 67.4% in 2024 (Figure 2.2.4). This steady climb signifies ongoing algorithmic refinement in object detection. Those capabilities are key for applications ranging from autonomous systems to medical image analysis, even as progress ramp flattens.

11. Data confirms scaling laws: More compute yields better performance

Chart showing the relationship between training compute and model performance
The report’s data consistently validates the scaling hypothesis central to modern AI development (See Figure 1.3.16, Page 58). Models trained with significantly more computational resources (petaFLOP) generally achieve higher performance across benchmarks. This trend, especially pronounced for LLMs, confirms that greater investment in training compute, coupled with larger datasets, remains a primary driver of capability improvements at the frontier, even as the resource requirements become astronomical. One notable example of this is Grok 3, a model trained using approximately 200,000 Nvidia H100 GPUs, representing a tenfold increase in compute over its predecessor, Grok 2. The compute led to significant improvement gains, exemplifying the scaling hypothesis in action. That is, allocating more computational power and data to model training continues to yield significant gains in AI performance.

12. US dominates global private AI investment, attracting $109B in 2024

Chart showing global private investment in AI by country
While China is catching up on the performance front, the financial engine of AI development remains overwhelmingly U.S.-centric. “‘The U.S. widens its lead in global AI private investment. U.S. private AI investment hit $109.1 billion in 2024, nearly 12 times higher than China’s $9.3 billion and 24 times the U.K.’s $4.5 billion.’ (Page 18, Highlight 3).” Generative AI was a major focus, attracting $33.9 billion globally in 2024 – an 18.7% increase from 2023 and over 8.5 times the 2022 level, representing over 20% of all AI private investment.

13. AI inference costs plummet over 280-fold in 18 months

Chart showing the decreasing cost of AI inference for fixed performance levels
The cost of using capable AI models has decreased dramatically. The report quantifies this with a striking example: “‘The cost of querying an AI model that scores the equivalent of GPT-3.5 (64.8) on MMLU… dropped from $20.00 per million tokens in November 2022 to just $0.07 per million tokens by October 2024… —a more than 280-fold reduction in approximately 18 months.’ (Page 29, Highlight 6).”

14. Training compute for frontier models doubles every five months

Chart showing the exponential growth of training compute for notable AI models
The computational appetite for training state-of-the-art AI continues its exponential ascent. “‘New research finds that the training compute for notable AI models doubles approximately every five months…’ (Page 28, Highlight 5).” This swift growth rate, far outpacing Moore’s Law.

15. Hardware energy efficiency improves 40% annually

Chart showing the improving energy efficiency of machine learning hardware
Amidst soaring compute demands, a counter-trend is the significant improvement in hardware energy efficiency. The report finds that ‘…energy efficiency has increased by 40% annually.’ (Page 29, Highlight 8).” For perspective, the Nvidia B100 (released March 2024) offers roughly 33.8 times the energy efficiency (16-bit FLOP/s per Watt) compared to the P100 (released April 2016) (Page 72).

16. Inference cost reductions show huge task-dependent variation

Chart showing varying rates of inference cost reduction for different LLM tasks
The dramatic fall in inference costs isn’t uniform across applications. “‘Depending on the task, LLM inference prices have fallen anywhere from 9 to 900 times per year.’ (Page 29, Highlight 6).” This wide variance underscores that efficiency gains are highly dependent on the specific AI task and the models optimized for it.

17. Open-weight models achieve near-parity with closed-weight leaders

Chart showing the narrowing performance gap between closed and open-weight models on Chatbot Arena
The performance gap between leading proprietary (closed-weight) and open-weight models has collapsed. “‘Last year’s AI Index revealed that leading open-weight models lagged significantly… By 2024, this gap had nearly disappeared… By February 2025, this gap had narrowed to 1.7%.’ (Page 14, Highlight 2).” Judging by human preference on Chatbot Arena, this near-parity demonstrates the increasing strength and viability of open approaches in achieving state-of-the-art results.

18. Hardware price-performance improves 30% annually

Chart showing the improving price-performance of machine learning hardware
The cost of AI computing power is consistently falling relative to its capability. “‘Price performance has improved, with costs dropping 30% per year…’ (Page 29, Highlight 8).” This means that for a fixed budget, R&D teams can access progressively more powerful hardware over time, further fueling AI development and deployment by making computation more financially accessible.

19. Speech recognition approaches human-level accuracy in benchmarks

Chart showing performance progress on the LibriSpeech automatic speech recognition benchmark
AI’s ability to transcribe human speech accurately continues to advance. On the LibriSpeech benchmark’s challenging test-clean subset, the Word Error Rate (WER) dropped to just 1.3% by late 2024 (Figure 2.2.14). This level of accuracy, nearing human performance even in less-than-ideal audio conditions, enables increasingly reliable and sophisticated voice-based interfaces and applications.

20. Llama 3.1 (405B) training emissions dwarf GPT-4’s, nearing 9,000 tons CO2e

Chart showing the increasing carbon emissions from training notable AI models
The environmental footprint of training frontier models is escalating dramatically. As the report highlights: “‘Carbon emissions from AI training are steadily increasing… GPT-3 (2020) at 588 tons, GPT-4 (2023) at 5,184 tons, and Llama 3.1 405B (2024) at 8,930 tons. For perspective, the average American emits 18 tons of carbon per year.’ (Page 29, Highlight 9).” Training Meta’s largest Llama 3.1 model generated roughly 72% more emissions than training GPT-4 and is equivalent to the annual emissions of nearly 500 average Americans.


Source and citation for the HAI 2025 AI Index Report:

Nestor Maslej, Loredana Fattorini, Raymond Perrault, Yolanda Gil, Vanessa Parli, Njenga Kariuki, Emily Capstick, Anka Reuel, Erik Brynjolfsson, John Etchemendy, Katrina Ligett, Terah Lyons, James Manyika, Juan Carlos Niebles, Yoav Shoham, Russell Wald, Tobi Walsh, Armin Hamrah, Lapo Santarlasci, Julia Betts Lotufo, Alexandra Rome, Andrew Shi, Sukrut Oak. “The AI Index 2025 Annual Report,” AI Index Steering Committee, Institute for Human-Centered AI, Stanford University, Stanford, CA, April 2025. The AI Index 2025 Annual Report by Stanford University is licensed under Attribution-NoDerivatives 4.0 International. Available at: https://aiindex.stanford.edu/report/

Related Articles Read More >

Abstract of modern high tech internet data center room with rows of racks with network and server hardware. 3d rendering
A startup says it found hidden memory behavior in NVIDIA GPUs and is building a security layer around it
Study shows LLMs can diagnose ER patients more accurately than physicians
These microrobots can collect nanoplastics from water
New Pistoia Alliance survey shows just 1% of professionals report AI’s value in the wet lab
rd newsletter
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, trends, and strategies in Research & Development.

R&D World Digital Issues

Fall 2025 issue

Browse the most current issue of R&D World and back issues in an easy to use high quality format. Clip, share and download with the leading R&D magazine today.

R&D 100 Awards
Research & Development World
  • Subscribe to R&D World Magazine
  • Sign up for R&D World’s newsletter
  • Contact Us
  • About Us
  • Drug Discovery & Development
  • Pharmaceutical Processing
  • Global Funding Forecast

Copyright © 2026 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search R&D World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • 2025 R&D 100 Award Winners
    • 2025 Professional Award Winners
    • 2025 Special Recognition Winners
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
  • Resources
    • Research Reports
    • Digital Issues
    • Educational Assets
    • Subscribe
    • Video
    • Webinars
    • Content submission guidelines for R&D World
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE