NEWS

OpenAI Big Upgrade: ChatGPT Can Now See, Hear, and Speak

In a couple of months, OpenAI’s chatbot ChatGPT has been receiving various new upgrades and enhancements on a regular basis. Recently, OpenAI has announced a series of new features, making the ChatGPT more intelligent. According to the official Blog post, OpenAi has announced that the company is starting to roll out new features and enhanced capabilities to the ChatGPT. 

OpenAI has made an official announcement regarding the rolling out process of the new Voice and image capabilities within its chatbot, ChatGPT. These latest features will offer a new, more innovative type of user interface by letting users have a voice conversation or show the chatbot what the users are talking about. 

The ChatGPT’s new Voice and imageability offers a more interactive way to use the chatbot in day-to-day life. For example, when taking a picture of a specific location while traveling, you can have a conversation about all the interesting information about it. Meanwhile, OpenAI is rolling out Voice and images in ChatGPT to Plus as well as the Enterprise users over the next two weeks. Alongside, Voice is rolling out on iOS and Android devices, while images will be available on all the platforms. 

ChatGpt is there to talk to you.

Users simply need to activate the ChatGPT along with the Voice prompts and can engage themselves in one-on-one conversations with the assistant. The new voice feature is packed with the latest text-to-speech functionality, which is highly capable of generating human-like audio just from text and a few seconds of demo speech by using the Whisper feature, which is OpenAI’s open-source recognition system that transcribes your spoken words into texts.  

Show ChatGPT an Image and let it answer.

With this latest update, users can now show the ChatGPT one or more than one image and let the ChatGPT answer the information regarding the image. In order to focus on a specific part of an object, you can use the drawing tool in the mobile app. The ChatGPT uses the multimodal GPT-3.5 and GPT-4 for understanding the image by also applying their certain language reasoning skills to a wide variety of images, including your photos, screenshots, and other documents that contain both text and images. 

Additionally, OpenAI has committed that it will gradually increase the availability of the voice and image tools and will be expanding the new Voice and image capabilities for the Plus and Enterprise users, including developers, so that they can experience the new enhanced interface very soon, in the next two weeks. 

ChatGPT Can Now See, Hear, and Speak

“If you like this article follow us on Google NewsFacebookTelegram, and Twitter. We will keep bringing you such articles.”

Related Articles

Back to top button