OpenAI is gearing up to launch its highly anticipated Voice Mode upgrade, powered by the advanced GPT-4o AI system, next week. This new feature promises to revolutionize spoken interactions with humanlike conversations.
The GPT-4o model, introduced in May, enhances the existing voice capabilities by providing real-time translations between speakers of different languages, allowing seamless communication across linguistic barriers. OpenAI showcased this groundbreaking feature in recent demos, highlighting its potential.
Currently, GPT-4o is available for text-based interactions, but the upgraded Voice Mode will be accessible to ChatGPT Plus subscribers for $20 a month. This “alpha” version is still in testing but will offer significantly improved performance compared to the previous GPT-4-powered voice mode.
OpenAI chief Sam Altman revealed that GPT-4o will greatly reduce the response delay, with audio inputs processed in as little as 232 milliseconds on average, compared to the 5.4 seconds delay experienced with GPT-4. The new system will also allow users to adjust the tone and pace of the voice and provide real-time language translation.
The upcoming Voice Mode aims to deliver a more natural and efficient conversational experience, bringing AI communication closer to human-like interaction.