In this pilot, ChatGPT Plus users will get to see first-hand how hyperrealistic OpenAI’s Advanced Voice Mode really is.
OpenAI began rolling out ChatGPT’s Advanced Voice Mode, giving its users their first access to GPT-4o’s hyperrealistic audio responses.
The alpha version will be available to a small group of ChatGPT Plus users, and OpenAI says the feature will gradually roll out to all Plus users in the last quarter of 2024. When OpenAI first showcased GPT-4o’s voice in May, the feature shocked audiences with quick responses and uncanny resemblance to a real human’s voice.
OpenAI says the video and screensharing capabilities showcased during its Spring Update will not be part of this alpha, launching at a later date. For now, the GPT-4o demo that blew everyone away is still just a demo, but some premium users will now have access to ChatGPT’s voice feature shown there.
ChatGPT’s old solution to audio used three separate models: one to convert your voice to text, GPT-4 to process your prompt, and then a third to convert ChatGPT’s text into voice. But GPT-4o is multimodal, capable of processing these tasks without the help of auxiliary models, creating significantly lower latency conversations. OpenAI also claims GPT-4o can sense emotional intonations in your voice, including sadness, excitement or singing.
In this pilot, ChatGPT Plus users will get to see first-hand how hyperrealistic OpenAI’s Advanced Voice Mode really is.
OpenAI says it’s releasing ChatGPT’s new voice gradually to closely monitor its usage. People in the alpha group will get an alert in the ChatGPT app, followed by an email with instructions on how to use it.