ChatGPT is about to become much more than a text-based chatbot and challenge traditional voice-based assistants like Siri and Alexa. Yesterday, OpenAI announced that it’s rolling out new voice and image capabilities to ChatGPT, which will allow users to talk with the chatbot and ask him questions about what they’re seeing.
The new voice and image capabilities will require a ChatGPT Plus or Enterprise subscription, and speaking with ChatGPT will only be possible on mobile. OpenAI has worked with voice actors to create five different voices for ChatGPT, and the new voice chat capability will use the same technology that Spotify uses to translate podcasts into other languages with an AI-generated voice that imitates the original content creator.
Sign up for our new free newsletter to get three time-saving tips each Friday — and get free copies of Paul Thurrott's Windows 11 and Windows 10 Field Guides (normally $9.99) as a special welcome gift!
"*" indicates required fields
OpenAI said that users will be able to engage in back-and-forth conversations with ChatGPT, and ask it to create content such as a bedtime story and more. Again, this sounds a lot like the upgraded Alexa with generative AI capabilities that Amazon demoed during its hardware and services event last week.
ChatGPT can now see, hear, and speak. Rolling out over next two weeks, Plus users will be able to have voice conversations with ChatGPT (iOS & Android) and to include images in conversations (all platforms). https://t.co/uNZjgbR5Bm pic.twitter.com/paG0hMshXb
— OpenAI (@OpenAI) September 25, 2023
With its new image capabilities, ChatGPT will also be able to analyze one or more images, including photos taken on your phone. On mobile, it will be possible to highlight a specific part of a photo to help ChatGPT focus on it. OpenAI detailed various use cases for these new image capabilities including data analysis, meal suggestions based on what you have in your fridge, or troubleshooting devices.
These new voice and image capabilities will roll out to ChatGPT Plus and Enterprise users over the next two weeks, with developers coming next. However, OpenAI said that at the moment, its language models are not optimized for non-English users.