ChatGPT Advanced Voice mode: release date, compatibility, and more

by Yaron

Advanced Voice Mode is a new feature for ChatGPT that enables users to hold real-time, humanlike conversations with the AI chatbot without the need for a text-based prompt window or back-and-forth audio. It was released in late July to select Plus subscribers after being first demoed at OpenAI’s Spring Update event.

When will I get Advanced Mode?

 
Introducing GPT-4o

Last week, OpenAI finally released its highly anticipated Advanced Voice feature to a select few of its ChatGPT-Plus subscribers. So, some people are already out there using it. We don’t know how large this initial rollout is, however, or how quickly OpenAI is inviting new members.

The company promises to roll it out to more users in the coming weeks, and to all Plus subscribers by this fall, although many users remain impatient to try it for themselves. So, the simple answer is that you’ll get access to it sometime between now and the fall, barring any further delays. “Fall” can be vague, but if we don’t see a full rollout to all users by December, that would constitute a delay.

Either way, you’ll know you have access when you receive an email inviting you to give it a try. You’ll also see a notification when you open the ChatGPT app.

What do I need to run Advanced Voice Mode on my phone?

the advanced voice alert on the chatgpt app

OpenAI

In addition to a Plus subscription, users will need an Android handset running app version 1.2024.206 or later, or an iPhone using iOS 16.4 or later and app version 1.2024.206 or later. Unfortunately, just having the right equipment isn’t enough to guarantee you a spot in the alpha release phase. What’s more, OpenAI has not released any details as to how or why it chooses the users it does.

If you do win the alpha release lottery and are selected, OpenAI will send you an email about it. You’ll also see a tooltip in the bottom-right corner of the ChatGPT mobile app that allows you to select the new feature.

Is OpenAI going to use my conversations to train its model?

Of course. Throughout the alpha release phase, OpenAI plans to leverage audio from conversations with advanced Voice Mode to train its models, assuming you haven’t yet turned off the app’s dat- sharing option.

Doing so is actually quite simple. On your mobile app, go to the Data Controls tab in your Settings menu and deselect Improve voice for everyone.

How often and how much can I use Advanced Voice Mode?

 
10 Minutes with ChatGPT’s Advanced Voice Mode

According to OpenAI, both the inputs and outputs for Advanced Voice come with daily usage limits, however, there’s no specifics on exactly how long those are and “precise limits are subject to change.” That said, user Himels Tech has already posted a video of themselves conversing with the AI for the better part of 10 minutes.

The AI will prompt users when they have 3 minutes of chatting left, before ending the conversation and sending the user back to the standard voice interface.

What can I, and can’t I, do with Advanced Voice Mode?

Khan!!!!!! pic.twitter.com/xQ8NdEojSX

— Ethan Sutin (@EthanSutin) July 30, 2024

Advanced Voice Mode is, at its core, simply a new way to interact with the same GPT-4o large language model that people already use for their text-based queries. So, in short, most anything you can do with ChatGPT, you can do with Advanced Voice, but with funny voices. From beatboxing to storytelling to counting really, really fast, early users are already putting the new feature through its paces.

There are safety guardrails and feature limits to what users can ask of the new mode, however. For one, users can’t use Advanced Voice to make new memories, nor can they use custom instructions or access GPTs using it. And while the AI will remember previous Advanced Voice conversations and be able to recall details of those talks, it cannot yet access previous chats conducted through the text prompt or the standard voice mode.

What’s more, Advanced Voice will not sing, no matter how you ask. Per the company, “to respect creators’ rights, we’ve put in place several mitigations, including new filters, to prevent advanced Voice Mode from responding with musical content including singing.”

Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
  • Computing

GPT-5: everything we know so far about OpenAI’s next frontier model
A MacBook Pro on a desk with ChatGPT's website showing on its display.

There’s perhaps no product more hotly anticipated in tech right now than GPT-5. Rumors about it have been circulating ever since the release of GPT-4, OpenAI’s groundbreaking foundational model that’s been the basis of everything the company has launched over the past year, such as GPT-4o, Advanced Voice Mode, and the OpenAI o1-preview.

Those are all interesting in their own right, but a true successor to GPT-4 is still yet to come. Now that it’s been over a year a half since GPT-4’s release, buzz around a next-gen model has never been stronger.
When will GPT-5 be released?
OpenAI has continued a rapid rate of progress on its LLMs. GPT-4 debuted on March 14, 2023, which came just four months after GPT-3.5 launched alongside ChatGPT. OpenAI has yet to set a specific release date for GPT-5, though rumors have circulated online that the new model could arrive as soon as late 2024.

Read more

  • Computing

OpenAI could release its next-generation model by December
ChatGPT giving a response about its knowledge cutoff.

OpenAI plans to release its next-generation frontier model, code-named Orion and rumored to actually be GPT-5, by December, according to an exclusive report from The Verge. However, OpenAI boss Sam Altman is already pushing back.

According to “sources familiar with the plan,” Orion will not initially be released to the general public, as the previous GPT-4 variants were. Instead, the company intends to hand the new model over to select businesses and partners, who will then use it as a platform to build their own products and services. This is the same strategy that Nvidia is pursuing with its NVLM 1.0 family of large language models (LLMs).

Read more

  • Computing

Radiohead’s Thom Yorke among thousands of artists who issue AI protest
Thom Yorke on stage.

Leading actors, authors, musicians, and novelists are among 11,500 artists to have put their name to a statement calling for a halt to the unlicensed use of creative works to train generative AI tools like OpenAI’s ChatGPT, describing it as a “threat” to the livelihoods of creators.

The open letter, comprising just 29 words, says: “The unlicensed use of creative works for training generative AI is a major, unjust threat to the livelihoods of the people behind those works, and must not be permitted.”

Related Posts

Leave a Comment