Behind millions of dollars of funding in AI sit enterprises with just a 5% average utilisation rate. Inference cost plus cost of ownership also rose to 41% from 34%

Reddit r/singularity News

Summary

Enterprises that rushed to buy massive GPU fleets for AI now face low utilization rates (5%) and rising costs (inference cost plus cost of ownership rose to 41% from 34%), highlighting significant infrastructure inefficiencies in AI deployment.

Well, Over the last few years after the Chat GPT rolled out, companies rushed to buy massive GPU fleets because AI demand exploded and compute was scarce but i think now it depends on more than just utilization like utilization, scheduling, inference efficiency, routing, governance, energy access, and operational management. The irony hits perfect, the technology designed to have the most efficient impact on human lives has this huge inefficiency of infrastructure problem Where majority budget goes out in figuring out allocation of hardware Source: [https://winbuzzer.com/2026/05/11/enterprises-face-underused-gpu-fleets-as-ai-costs-rise-xcxwbn](https://winbuzzer.com/2026/05/11/enterprises-face-underused-gpu-fleets-as-ai-costs-rise-xcxwbn)
Original Article

Similar Articles

AI and compute

OpenAI Blog

OpenAI releases an analysis demonstrating that compute used in largest AI training runs has grown exponentially at a 3.4-month doubling time since 2012, representing a 300,000x increase and vastly outpacing Moore's Law. The analysis suggests this trend will likely continue and calls for increased academic AI research funding to address rising computational costs.

The state of enterprise AI

OpenAI Blog

OpenAI released a 2025 report on enterprise AI adoption, revealing that over 1 million business customers use OpenAI tools with significant scaling metrics: ChatGPT message volume grew 8x and API reasoning token consumption increased 320x year-over-year. The report highlights measurable productivity gains (40-60 minutes saved daily), global expansion, and an emerging gap between AI leaders and laggards in organizations.

The Inference Shift (8 minute read)

TLDR AI

This article analyzes Cerebras' upcoming IPO as a signal of the 'inference shift' in AI hardware, arguing that while Nvidia dominates GPU-based training, the future of AI compute is becoming increasingly heterogeneous to support inference workloads.

AI inference just plays by different rules (9 minute read)

TLDR AI

The article argues that AI inference poses unique challenges to cloud data infrastructure, likening its demand to high-concurrency OLTP systems rather than traditional human-speed applications. It emphasizes the need to optimize storage and data access layers to handle the 'AI data tsunami' driven by autonomous agents.