Behind millions of dollars of funding in AI sit enterprises with just a 5% average utilisation rate. Inference cost plus cost of ownership also rose to 41% from 34%
Summary
Enterprises that rushed to buy massive GPU fleets for AI now face low utilization rates (5%) and rising costs (inference cost plus cost of ownership rose to 41% from 34%), highlighting significant infrastructure inefficiencies in AI deployment.
Similar Articles
AI and compute
OpenAI releases an analysis demonstrating that compute used in largest AI training runs has grown exponentially at a 3.4-month doubling time since 2012, representing a 300,000x increase and vastly outpacing Moore's Law. The analysis suggests this trend will likely continue and calls for increased academic AI research funding to address rising computational costs.
The state of enterprise AI
OpenAI released a 2025 report on enterprise AI adoption, revealing that over 1 million business customers use OpenAI tools with significant scaling metrics: ChatGPT message volume grew 8x and API reasoning token consumption increased 320x year-over-year. The report highlights measurable productivity gains (40-60 minutes saved daily), global expansion, and an emerging gap between AI leaders and laggards in organizations.
The Inference Shift (8 minute read)
This article analyzes Cerebras' upcoming IPO as a signal of the 'inference shift' in AI hardware, arguing that while Nvidia dominates GPU-based training, the future of AI compute is becoming increasingly heterogeneous to support inference workloads.
AI inference just plays by different rules (9 minute read)
The article argues that AI inference poses unique challenges to cloud data infrastructure, likening its demand to high-concurrency OLTP systems rather than traditional human-speed applications. It emphasizes the need to optimize storage and data access layers to handle the 'AI data tsunami' driven by autonomous agents.
OpenAI's deployment company move says more about the AI gap than any benchmark[D]
OpenAI launched a deployment company with $4B investment, acquiring Tomoro, to embed engineers into enterprises for AI deployment, signaling that the gap between model capability and production deployment is widening.