Skip to main content

ChatGPT Advanced Voice mode: release date, compatibility, and more

Advanced Voice Mode is a new feature for ChatGPT that enables users to hold real-time, humanlike conversations with the AI chatbot without the need for a text-based prompt window or back-and-forth audio. It was released in late July to select Plus subscribers after being first demoed at OpenAI’s Spring Update event.

According to the company, the feature “offers more natural, real-time conversations, allows you to interrupt at any time, and senses and responds to your emotions.” It can even take breath breaks and simulate human laughter during conversation. The best part is that access is coming soon, if you don’t have it already.

When will I get Advanced Mode?

Introducing GPT-4o

Last week, OpenAI finally released its highly anticipated Advanced Voice feature to a select few of its ChatGPT-Plus subscribers. So, some people are already out there using it. We don’t know how large this initial rollout is, however, or how quickly OpenAI is inviting new members.

The company promises to roll it out to more users in the coming weeks, and to all Plus subscribers by this fall, although many users remain impatient to try it for themselves. So, the simple answer is that you’ll get access to it sometime between now and the fall, barring any further delays. “Fall” can be vague, but if we don’t see a full rollout to all users by December, that would constitute a delay.

Either way, you’ll know you have access when you receive an email inviting you to give it a try. You’ll also see a notification when you open the ChatGPT app.

What do I need to run Advanced Voice Mode on my phone?

the advanced voice alert on the chatgpt app
OpenAI

In addition to a Plus subscription, users will need an Android handset running app version 1.2024.206 or later, or an iPhone using iOS 16.4 or later and app version 1.2024.206 or later. Unfortunately, just having the right equipment isn’t enough to guarantee you a spot in the alpha release phase. What’s more, OpenAI has not released any details as to how or why it chooses the users it does.

If you do win the alpha release lottery and are selected, OpenAI will send you an email about it. You’ll also see a tooltip in the bottom-right corner of the ChatGPT mobile app that allows you to select the new feature.

Is OpenAI going to use my conversations to train its model?

Of course. Throughout the alpha release phase, OpenAI plans to leverage audio from conversations with advanced Voice Mode to train its models, assuming you haven’t yet turned off the app’s dat- sharing option.

Doing so is actually quite simple. On your mobile app, go to the Data Controls tab in your Settings menu and deselect Improve voice for everyone.

How often and how much can I use Advanced Voice Mode?

10 Minutes with ChatGPT's Advanced Voice Mode

According to OpenAI, both the inputs and outputs for Advanced Voice come with daily usage limits, however, there’s no specifics on exactly how long those are and “precise limits are subject to change.” That said, user Himels Tech has already posted a video of themselves conversing with the AI for the better part of 10 minutes.

The AI will prompt users when they have 3 minutes of chatting left, before ending the conversation and sending the user back to the standard voice interface.

What can I, and can’t I, do with Advanced Voice Mode?

Khan!!!!!! pic.twitter.com/xQ8NdEojSX

— Ethan Sutin (@EthanSutin) July 30, 2024

Advanced Voice Mode is, at its core, simply a new way to interact with the same GPT-4o large language model that people already use for their text-based queries. So, in short, most anything you can do with ChatGPT, you can do with Advanced Voice, but with funny voices. From beatboxing to storytelling to counting really, really fast, early users are already putting the new feature through its paces.

There are safety guardrails and feature limits to what users can ask of the new mode, however. For one, users can’t use Advanced Voice to make new memories, nor can they use custom instructions or access GPTs using it. And while the AI will remember previous Advanced Voice conversations and be able to recall details of those talks, it cannot yet access previous chats conducted through the text prompt or the standard voice mode.

What’s more, Advanced Voice will not sing, no matter how you ask. Per the company, “to respect creators’ rights, we’ve put in place several mitigations, including new filters, to prevent advanced Voice Mode from responding with musical content including singing.”

Andrew Tarantola
Andrew has spent more than a decade reporting on emerging technologies ranging from robotics and machine learning to space…
GPT-4: everything you need to know about ChatGPT’s standard AI model
A laptop opened to the ChatGPT website.

People were in awe when ChatGPT came out, impressed by its natural language abilities as an AI chatbot originally powered by the GPT-3.5 large language model. But when the highly anticipated GPT-4 large language model came out, it blew the lid off what we thought was possible with AI, with some calling it the early glimpses of AGI (artificial general intelligence).
What is GPT-4?
GPT-4 is the newest language model created by OpenAI that can generate text that is similar to human speech. It advances the technology used by ChatGPT, which was previously based on GPT-3.5 but has since been updated. GPT is the acronym for Generative Pre-trained Transformer, a deep learning technology that uses artificial neural networks to write like a human.

According to OpenAI, this next-generation language model is more advanced than ChatGPT in three key areas: creativity, visual input, and longer context. In terms of creativity, OpenAI says GPT-4 is much better at both creating and collaborating with users on creative projects. Examples of these include music, screenplays, technical writing, and even "learning a user's writing style."

Read more
OpenAI just took the shackles off the free version of ChatGPT
ChatGPT results on an iPhone.

OpenAI announced the release of its newest snack-sized generative model, dubbed GPT-4o mini, which is both less resource intensive and cheaper to operate than its standard GPT-4o model, allowing developers to integrate the AI technology into a far wider range of products.

It's a big upgrade for developers and apps, but it also expands the capabilities and reduces limitations on the free version of ChatGPT. GPT-4o mini is now available to users on the Free, Plus, and Team tiers through the ChatGPT web and app for users and developers starting today, while ChatGPT Enterprise subscribers will gain access next week. GPT-4o mini will replace the company's existing small model, GPT-3.5 Turbo, for end users beginning today.

Read more
ChatGPT may have more paid subscribers than this popular streaming service
OpenAI CEO Sam Altman standing on stage at a product event.

OpenAI's steamrolling of its rivals continued apace this week, and a new study estimates just how much success it's had in winning over paid subscribers through ChatGPT Plus.

According to a report published by Futuresearch this week, OpenAI's products are far and away the most popular -- and profitable -- in the AI space. Per the study, OpenAI has an estimated annual recurring revenue of $3.4 billion dollars.

Read more