OpenAI has officially begun the rollout of its new Advanced Voice Mode for ChatGPT, which introduces hyper-realistic audio responses. Initially available to a limited number of ChatGPT Plus users, the feature is set to expand to all Plus subscribers by fall 2024.
Powered by the innovative GPT-4o model, the Advanced Voice Mode integrates voice recognition, text processing, and speech generation into a single system. This technological advancement allows for quicker responses and enables the AI to interpret emotional cues in users’ voices, such as excitement or sadness.
The feature currently offers four distinct voices—Juniper, Breeze, Cove, and Ember—developed in collaboration with professional voice actors. This approach aims to prevent the AI from mimicking real individuals, addressing concerns related to deepfake technology.
Additionally, OpenAI has implemented new filters to block requests for generating copyrighted audio and to ensure compliance with legal standards, especially in light of recent controversies surrounding AI and copyright infringement. Prior to the launch, OpenAI conducted extensive testing with over 100 external testers who speak 45 different languages, refining the model’s capabilities and addressing potential vulnerabilities. A report detailing these safety measures is expected to be released soon.
The Advanced Voice Mode represents a significant step forward in enhancing user interaction with AI, making conversations more engaging and lifelike.