ChatGPT voice mode is a weaker model

Simon Willison's Blog News

Summary

ChatGPT's voice mode runs on a weaker GPT-4o era model with an April 2024 knowledge cutoff, significantly older than OpenAI's latest capabilities. The article highlights a growing gap between OpenAI's consumer voice interface and its more advanced paid models, driven by differences in reward signal clarity and B2B market incentives.

No content available
Original Article Export to Word Export to PDF
View Cached Full Text

Cached at: 04/20/26, 08:28 AM

# ChatGPT voice mode is a weaker model Source: https://simonwillison.net/2026/Apr/10/voice-mode-is-weaker/ 10th April 2026 I think it's non-obvious to many people that the OpenAI voice mode runs on a much older, much weaker model - it feels like the AI that you can talk to should be the smartest AI but it really isn't. If you ask ChatGPT voice mode for its knowledge cutoff date it tells you April 2024 - it's a GPT-4o era model. This thought inspired by [this Andrej Karpathy tweet](https://twitter.com/karpathy/status/2042334451611693415) about the growing gap in understanding of AI capability based on the access points and domains people are using the models with: > [...] It really is simultaneously the case that OpenAI's free and I think slightly orphaned (?) "Advanced Voice Mode" will fumble the dumbest questions in your Instagram's reels and *at the same time*, OpenAI's highest-tier and paid Codex model will go off for 1 hour to coherently restructure an entire code base, or find and exploit vulnerabilities in computer systems. This part really works and has made dramatic strides because 2 properties: 1. these domains offer explicit reward functions that are verifiable meaning they are easily amenable to reinforcement learning training (e.g. unit tests passed yes or no, in contrast to writing, which is much harder to explicitly judge), but also 2. they are a lot more valuable in b2b settings, meaning that the biggest fraction of the team is focused on improving them.

Similar Articles

ChatGPT can now see, hear, and speak

OpenAI Blog

OpenAI is rolling out new voice and image capabilities to ChatGPT Plus and Enterprise users, enabling users to have voice conversations and share images for multimodal interactions powered by GPT-3.5/GPT-4 and custom text-to-speech models.

Introducing ChatGPT Pro

OpenAI Blog

OpenAI launches ChatGPT Pro, a $200/month subscription plan offering unlimited access to advanced models including o1, o1-mini, GPT-4o, and Advanced Voice, plus o1 pro mode for compute-intensive reasoning tasks.

How the voices for ChatGPT were chosen

OpenAI Blog

OpenAI explains its process for selecting five distinct voices for ChatGPT's Voice Mode feature, involving professional voice actors, casting directors, and a five-month selection process. The company addresses controversy over the 'Sky' voice, clarifying it is not an imitation of Scarlett Johansson and was cast before any outreach to her.

Introducing ChatGPT

OpenAI Blog

OpenAI introduces ChatGPT, a conversational AI model fine-tuned from GPT-3.5 using reinforcement learning from human feedback (RLHF). The model is designed to answer follow-up questions, admit mistakes, and reject inappropriate requests, with free access provided during the research preview.

Introducing GPT-4o and more tools to ChatGPT free users

OpenAI Blog

OpenAI is launching GPT-4o, its newest flagship model with improved text, voice, and vision capabilities, and making it available to free ChatGPT users with usage limits. They are also releasing a new macOS desktop app and redesigned UI for ChatGPT.