That’s why it’s always interesting to see new updates and features added. However, the latest update is particularly thought-provoking.
We know that progress toward general AI (AI that can do anything) is on the part of developers. OpenAI plans. And to achieve that ambitious goal, vision and the ability to speak and hear are critical. So it’s only natural that these features would be added to ChatGPT as it celebrates its one-year anniversary since it became available to everyone.
But how does this affect our relationships with our new digital friends? What does it add to its ability to integrate into our lives and help us solve everyday challenges? And what does it mean for the larger question of ChatGPT’s (and AI in general’s) place in society?
What is the new ChatGPT update?
First, ChatGPT gains the ability to see effectively by gaining the ability to analyze and extract linguistic information from images. Not only can you upload a photo and have it explain what you see there, but you can also use it to perform more complex tasks, such as diagnosing how to fix a broken machine like a bicycle or lawn mower. You may also be able to use .
This means that, in theory, ChatGPT could analyze not just photos, but also charts, visualizations, handwriting, and all kinds of unstructured data around us.
It’s obvious that it has a variety of everyday uses, from creating a Facebook Marketplace listing for the products you want to sell to turning whiteboard doodles into easy-to-read notes.
Examples of uses OpenAI itself has cited include snapping items in the fridge, asking what you can make for dinner, and having live conversations about photos.
by new york timesHowever, there are limitations to the capabilities deployed in ChatGPT, some by design, for example, limiting the methods that can be used to analyze human faces. This is done to prevent it from being used to violate privacy, and is consistent with how OpenAI has previously restricted its products.
Another change that could have far-reaching implications is that ChatGPT can now speak and listen, at least when using the mobile app.
That’s why the voice revolution has made it commonplace these days to talk to machines like Siri and Alexa. And we’re all pretty used to the fact that they can only respond to us in a limited number of ways, and generally speaking, the most helpful thing they can do is turn other devices on and off. It’s about switching.
That’s why talking to ChatGPT is such an interesting idea. ChatGPT has the potential to make conversations so much more natural and fluid that it creates the illusion that you’re talking to a real person.
In addition to having back-and-forth conversations, it can also simulate audible voices, which can be used, for example, to voice an AI avatar.as well as many other more sinister purposes.
I say “potential” because I tried it right after it was released and I have to say it’s not quite there yet.
ChatGPT had some trouble understanding what I was saying. And what’s even weirder is that the training data to let you know it can speak doesn’t seem to be updated. When I used the voice feature to ask for help, I was flatly told (in a fully synthesized human voice) that there was no voice feature. I’ve also seen reports of people having difficulty understanding different accents and dialects.
What does this mean?
The transition to an era in which machines can not only think but also see, hear, and speak is clearly of considerable importance. I think there are many people who would be tempted to say that none of these things can be done well yet. But it’s clear that things are just getting started.
Visual and audio capabilities mean that ChatGPT technology is likely to appear in increasingly portable technologies.we already have glasses Overlay ChatGPT You’ll never get stuck trying to find a solution to a problem because it’s right in front of you.And these will help you create casual chat.
Some of them can already hear and speak, but these use third-party extensions to add functionality, and OpenAI’s own integration technology makes it (potentially) much smoother. It should provide a great experience.
But real-time, AI-powered image analysis, instantly available anywhere, has the potential to be truly transformative in many fields.
However, some ethical considerations arise. Most pressingly, although ChatGPT was released with various restrictions on its operation intended to prevent its use for unethical purposes, these were quickly circumvented and in some cases completely removed. It’s worth bearing in mind that it was removed in .
If this is done using ChatGPT’s visual capabilities, the potential impact could be even more severe, especially if an unethical actor finds a way to circumvent the facial recognition block.
Also, shortly after ChatGPT was released, copycat versions started appearing that worked without restrictions. In some cases, they were sold by their creators as tools specifically intended to break the law. Would the same thing happen with the ability to imitate vision or voice? I think it would be pretty foolish to think it wouldn’t.
Exploration of general artificial intelligence
But there may be something more to worry us about.
With the latest updates, ChatGPT is becoming increasingly multimodal. This means that it can understand and interact with various forms of input, including not only words but also images and audio.
This is important because the goal of AI development is necessarily artificial general intelligence (AGI). This is a term used to describe machines that, like humans, can perform any task as long as they have the necessary data. Becoming multimodal can easily be described as a big step towards this.
It is no exaggeration to say that AGI is still a long way off. Google Engineering Director Ray Kurzweil estimates we’ll get there around 2045, and DeepMind CEO Demis Hassabis believes it will happen within the next few decades.
However, philosopher Nick Bostrom believe “Superintelligence” will arrive early in the next century. Professor Stuart Russell is an AI pioneer and founder of the Center for Human-Compatible AI. To tell That is still a long way off, and there are still big questions to be solved.
When AGI does arrive, it could have quite a big impact.On the question of whether we will live a life of luxury while machines create everything we need, or whether we will become far removed from it. dark fateopinions are still divided.
What should I do then?
For better or worse, governments and those in positions of decision-making do not seem to be listening to the advice of those who signed this agreement. Pause a giant AI experiment Plea.
This means that the development of AI is likely to continue and accelerate. Features like the one added by ChatGPT will become part of our daily lives. It also provides a more reliable, more powerful, and continually improving user experience. This means more applications and devices will incorporate the technology, and an ever-growing list of social and industrial use cases.
Since ChatGPT can now be seen and heard, I don’t think it will be long before someone figures out how to touch, smell, and taste it. It would then be equipped with all the same sensory capabilities that we have, and theoretically be able to fully understand how we perceive our environment.
This gives us the potential to help us discover so many insights, including information about the world and our interactions with it, that our organic brains are beyond. Then, based on our understanding of our own capabilities, they provide us with that information in a way that they know we can use.
Thanks to AI, we are on a journey where the destination is never clear. But the kind of advances that have been made since ChatGPT was launched (less than a year) have made two things clear. It’s picking up speed and you can only guess what surprises will be around the next corner.