Join us as we return to New York on June 5 to collaborate with leaders to explore comprehensive methods for auditing AI models for bias, performance, and ethical compliance in diverse organizations. Find out how you can attend here.
Sometimes it feels like the world is suddenly moving forward. That’s certainly how it felt 18 months ago when OpenAI first launched ChatGPT. According to a UBS study, ChatGPT reached over 100 million users in the first two months of its release, making it the fastest growing consumer app in history. At the time, there was a palpable feeling that something dramatic was happening.
Then the world shifted from fascination and excitement to fear and anxiety about the potential for imminent extermination of malicious AIs. Remember the “break» proposed by Elon Musk and several hundred others, arguing that AI development should stop for six months? This did not happen, although Musk – ironically – began creating his own ChatGPT-like alternative during this time.
Some, like AI researcher Eliezer Yudkowsky, worried that ChatGPT pointed to a near future where AI would “achieve intelligence more intelligent than humans” and kill everyone. It even became fashionable for people to rate their belief in the probability of this outcome, called “probability of doom” or “P(doom)”. “What’s your P-(doom)?” » was the subject of dinner conversationsat least in Silicon Valley and other high-tech hangouts.
People who claimed high P(doom) were labeled “doomers.” This very includedded Geoffrey Hinton, one of the “godfathers of AI”, who decided a year ago that the AI technology he helped advance was advancing much faster than expected. His concerns about potential existential risks from increasingly powerful AI systems prompted him to change his stance.
Paradigm shift
Nevertheless, we are still here. Maybe the risks are still there too – they will just exist in the future. But instead of The terminatorwe have friendly virtual assistants like samanotThat. This new wizard ability is based on GPT-4o (“o” for omni)released this week by OpenAI, which can “reason through audio, vision and text in real time.” It can converse using natural speech without noticeable lag, read emotional signals (and respond in kind), and process visual input. In line with these new capabilities, Senior Editor Mark Sullivan of Fast business concluded that ChatGPT “becomes more emotionally intelligent.”
Ryan Morrison, AI editor at Tom's guide laid: “What we are seeing – and other companies will catch up – is the dawn of a new era in human-machine interface technology. Morrison added: “This is as significant a paradigm shift in access to information as the first printing press, the typewriter, the personal computer, the Internet or even the smartphone. » I had to agree with this observation after watching the live demo of GPT-4o. This indeed felt like another significant leap forward.
AGI or narrow AI?
And yet, this really shouldn't be happening, at least not right now. Only a few months ago it was believed that this level of achievement would require artificial general intelligence (AGI), requiring significant scientific advances in areas such as natural language processing, emotional intelligence, and creativity.
No one is claiming that this new level of virtual assistant is AGI, general intelligence that matches or exceeds human performance across a wide range of cognitive tasks. I asked ChatGPT about this and he agreed that GPT-4o is not yet AGI. It said: “GPT-4o should be described as an advanced form of weak (or narrow) AI. It is a sophisticated tool that combines multiple types of data processing in innovative ways, pushing the boundaries of what AI can do within specified parameters, while operating under the limits of a designed machine for particular applications, rather than an entirely autonomous and versatile intellect. .”
Salon tip or useful for applications?
THE GPT-4o demo it was like magic, but will it translate into any real use? Adding human-like interaction will make AI technologies more natural and engaging to the user and could clearly improve user experience in various applications including customer service, virtual assistance, education and entertainment.
Kevin Roose, tech columnist at The New York Timesrecently spoken on the Hard Fork podcast and in the Times about the “AI friends” he created across multiple platforms, including Nomi, Kindroid, and Replika. All of these companies use Large Language Model (LLM) technology, like that which underpinned ChatGPT before GPT-4o. Roose built 18 custom companions, gave each one a story, and was able to chat with them by texting back and forth. He could also speak with some of them, but with a significant delay. Even before the announcement of GPT-4o and its ability to “see” and communicate emotions, and with near real-time voice responses, he concluded: “The technology necessary for realistic AI companionship is already there. »
Other companies already rely on GPT-4o, including one that adds a face to voice. With synchronized expressions and lip movements, this capability will add to the already notable appeal of AI Friends, but will also find applications in healthcare, grooming, coaching and education – where a face friendly can make the difference. Adding a human-like face makes it easier to imagine virtual teachers, health advisors, or brand representatives.
Greater ability to integrate audio, visual and text data opens up potentially transformative opportunities in industries like healthcare, where it could lead to more nuanced patient interaction and personalized care plans.
Doctors are already exploring the possible impact of this advancement, including virtual health assistants functioning as “24/7 health advisors, providing immediate medical advice, reminding patients about medications and even offering support in matters mental health”. This could also include “empathetic support for mental health concerns,” which complements traditional therapies by recognizing and responding to subtle cues in a patient's speech or behavior.
In education, this technology could cater to varied learning styles with adaptive content responding in real time to the needs of each student. For example, in this videoSal Kahn – the founder of Khan Academy whose mission is to provide free education – and his son Imran work with GPT-4o to solve a math problem.
As GPT-4o and similar offerings continue to evolve, we can anticipate more sophisticated forms of natural language understanding and emotional intelligence. This could lead to AI that not only understands complex human emotions, but also responds in increasingly appropriate and helpful ways. The future could see AI becoming an integral part of emotional support networks, providing truly empathetic and informed companionship and help.
AI's journey from a niche technology to a fundamental part of our daily interactions is both exhilarating and daunting. To responsibly navigate this AI revolution, it is essential that developers, users, and policymakers engage in rigorous and ongoing dialogue about the ethical use of these technologies. As GPT-4o and similar AI tools become increasingly integrated into our daily lives, we must navigate this transformation journey with wisdom and foresight, ensuring that AI remains a tool that empowers rather than does not diminish our humanity.
Gary Grossman is senior vice president of the technology practice at Edelmann and Global Head of the Edelman AI Center of Excellence.
DataDecisionMakers
Welcome to the VentureBeat community!
DataDecisionMakers is where experts, including data technicians, can share data insights and innovations.
If you want to learn more about cutting-edge ideas and up-to-date information, best practices, and the future of data and data technology, join us at DataDecisionMakers.
You might even consider contribute to an article your own!