What to Learn
- OpenAI has released a new version of its robot that users can talk to.
- ChatGPT can now understand people’s words, talk back with a computer voice, and look at pictures.
- The robot could make technology more open to everyone with its new skills.
ChatGpt New Update
You can now talk to your robot, which could change how we use artificial intelligence (AI).
OpenAI has released a new version of its robot that can talk to people. ChatGPT can now “see, hear, and speak.” The bot can hear spoken language, talk back with a fake voice, and look at pictures.
Raffi Kryszek, the AI lead at Proto Inc., told Lifewire via email, “Using spoken words to interact with AI chatbots gives a sense of natural communication and caters to our natural preference for verbal exchange.” “This way of communicating is often faster than typing, making things easier, especially on devices or places where typing isn’t possible.”
Chatting With ChatGpt Bot
The latest update to the chatbot, the biggest one from OpenAI since GPT-4, lets people use the ChatGPT mobile tool to talk to the chatbot. The chatbot has five different robot sounds that users can choose from. They can also use a feature called GPT-4-Vision to show pictures to ChatGPT and point out places to look at or talk about.
The company’s website says, “Take a picture of a landmark while traveling and have a live conversation about what’s interesting about it.” “When you get home, take pictures of your fridge and pantry to figure out what to make for dinner. If you want a step-by-step plan, ask follow-up questions. After dinner, you can help your child solve a math problem by taking a picture of the problem, circling it, and having the computer tell you both what to do.
ChatGPT can imitate voices and make speech after hearing just a bit of someone speaking because it can understand voices. The updated voice feature in ChatGPT can now tell bedtime stories, help settle dinner table arguments, and read out loud what users have typed.
OpenAI has admitted that this feature could be used to trick people or commit fraud. Even though these worries exist, the company said that ChatGPT will only use sounds already in the system and have been approved by the company in advance.
Chris Callison-Burch, a professor of computer and information science at the University of Pennsylvania, said in an email that newer chatbots like OpenAI’s ChatGPT are much better at having conversations and understanding what users want than the older versions of Alexa, Siri, and Google Assistant. “As soon as smart assistants use generative AI, I expect them to leap forward.”
The Future of Artificial Intelligence
Close on the heels of the release of DALL-E 3, OpenAI’s newest and most advanced picture generation system, ChatGPT’s multimodal features were improved.
OpenAI says that DALL-E 3 also has natural language processing. This means that users can talk to the model to change the results and work with ChatGPT to help make picture prompts.
Kryszek said that voice-activated AI chatbots will be able to understand different accents and languages in the not-too-distant future. This will make technology more open and global.
“This change will be paired with the ability to read emotions from our voices,” he said. “This will make digital assistants more empathetic.” “These innovations are likely to affect every part of our lives, from gadgets to cars, with strong voice biometrics ensuring the highest level of security. As these systems improve, we’ll see a mix of voice, visual, and tactile feedback, bringing in a new era of immersive, multi-dimensional digital interactions.