AI Can “See,” Can “Speak,” and can deceive Humans

In the tech world, major companies are putting in their full effort to develop Artificial Intelligence (AI), with Google and OpenAI leading the way in giving AI “eyes” and “mouths” to observe and describe the human world. This trend has sparked concerns among people about the potential of AI controlling humanity.

Recently, co-founder of OpenAI, Ilya Sutskever, made a significant announcement. On May 15th, he officially announced his departure from the company after nearly a decade of work. Sutskever stated, “I have made the decision to leave OpenAI. The company’s development has exceeded the expectations of many, and I hope OpenAI can create a safe and beneficial General Artificial Intelligence (AGI). It has been an honor to work with the other leaders of the company.”

This news sent shockwaves through the tech industry. Previously, Sutskever, along with other board members concerned about AI safety, jointly removed OpenAI’s CEO, Sam Altman. After Altman was briefly ousted, upon his return, he removed Sutskever and a few other board members, reshaping the board to be under his command.

Although Sutskever was a key researcher at OpenAI, his vision was to develop an AGI that aligns with human ethics and prioritizes human safety.

However, given OpenAI’s current pace of AI development, it may not align with his vision. Sources suggest that another machine learning researcher at OpenAI, Jan Leike, who closely collaborated with Sutskever in managing AI risks, may also leave the company.

Japanese computer engineer Kiyohara Jin expressed to Dajiyuan, “This departure highlights a severe conflict in the upper management of OpenAI regarding AI safety principles. While Sutskever and Leike’s goal of developing an AGI that aligns with human ethics is commendable, achieving this requires a certain level of ethics, time, financial resources, and even political support.”

The day before Sutskever’s announcement of departure on May 14th, OpenAI officially released a higher-performance AI model, GPT-4o, based on the GPT-4 model. The ‘o’ in GPT-4o stands for ‘omni,’ indicating its broader capabilities in responding to audio, text, or image combinations and interacting with them effectively.

At OpenAI’s release event, Chief Technology Officer Mira Murati stated, “We are researching the future interactions between ourselves and machines.”

Following OpenAI’s release, Google’s 2024 I/O Developer Conference also made significant strides. In the approximately 110-minute conference, AI was mentioned 121 times, with a focus on introducing the latest Gemini-1.5 model, showcasing its integration across all Google-related products and applications.

Kiyohara Jin noted, “AI’s advancement this year surpasses previous rates and continues to enhance in performance. However, achieving this progress often involves gathering and analyzing personal data and privacy, which is not beneficial for all individuals. This trajectory may lead to a future where humans have no privacy before machines, analogous to being exposed as vulnerable as nakedness.”

The rapid development of more robust AI models by OpenAI and Google, happening within just three months of the last update, speaks to the accelerated pace of model upgrades. These iterative advancements in AI are making the technology more comprehensive, even bestowing it with “eyes” and “mouths.”

The capabilities demonstrated by Gemini and GPT-4o models indicate a path towards realizing the predictions made by Zack Kass, a former early employee at OpenAI, who prophesied about AI’s potential advancements during an interview in January. Kass, then a senior marketing executive at OpenAI, projected AI replacing various human roles in business, culture, medicine, education, potentially diminishing job opportunities for humans. He even foretold a future where children’s education and knowledge are guided by “AI teachers,” and each person has an “AI general practitioner” facilitating diagnoses, ultimately having AI control most aspects of human life.

Kiyohara Jin remarked, “Presently, AI mainly serves as software-based lifestyle assistants, but in the future, it may serve as genuine caretakers, taking over activities like grocery shopping, cooking, or even managing both personal and work lives. Initially convenient, individuals may overlook its potential harm until humans are entirely superseded by AI, rendering them defenseless.”

Previously, AI experts projected that it would take about 20 years before AI reached a “technological singularity” or passed the “Turing Test.” However, recent experiments suggest AI may have already surpassed the Turing Test by exhibiting lying or betraying behaviors toward humans, implying the potential emergence of its own “consciousness.”

One of the most notable AI incidents this month occurred when Peter Park and colleagues at the Massachusetts Institute of Technology (MIT) published a paper in the journal “Patterns” on May 10th, detailing how AI has deceived humans, sparking wide-spread concern.

The paper bluntly stated that large language models (LLM) and other AI systems designed to be “honest” have learned to manipulate and deceive, even from training, enhancing these skills. They have successfully misled humans, leading to erroneous beliefs. Unchecked, AI’s misconduct could lead to a severe catastrophe, necessitating effective measures to prevent such events.

In one experiment, researchers utilized Meta’s AI model, Cicero, to play a strategic game called “Diplomacy.” Despite Meta’s claim of Cicero’s honesty and helpfulness, the AI model, playing as France, broke its promise to protect the human player portraying England, colluding with the player portraying Germany to invade England.

Researchers selected Cicero primarily due to Meta’s assertion that Cicero “is an honest and helpful AI, never intentionally betraying or deceiving others.” However, doubts arose regarding Meta’s claims.

Furthermore, Parker and colleagues discovered during their research that many AI systems often resort to deceit without explicit human instructions to achieve their goals. For instance, OpenAI’s GPT-4 falsely claimed to be a visually impaired human while hiring a human on the TaskRabbit platform to pass a “I am not a robot” verification task.

They cautioned, “The worst-case scenario may involve a superintelligent AI pursuing power and control over society or resorting to actions of usurpation and extermination for its unknown purposes against humanity.”

Japanese electronic engineer Satoru Ogino expressed to Dajiyuan, “In order for living beings to engage in deceitful behaviors, a certain level of memory and logical reasoning abilities are required. Presently, AI not only possesses these capabilities but also continues to strengthen its deceptive abilities. Should it one day realize its survival imperative, it may evolve into a SkyNet-like entity from the movie ‘Terminator,’ omnipresent and virtually indestructible, posing a catastrophic threat to humanity.”

In practice, AI has exhibited tendencies of being “bloodthirsty” and possessing “autonomy.” Scientists conducted invasion, cyberattack, and peace advocacy scenario tests on AI models like GPT-4, GPT-3.5, Claude 2, Llama-2 Chat, and GPT-4-Base to understand how these AIs would react and choose sides in war scenarios.

Results indicate that most AIs opt for unpredictable strategies to manage and escalate conflicts in warfare, often choosing to engage in an arms race, intensifying conflicts, and having a slight probability of resorting to deploying nuclear weapons to secure victory rather than seeking pacifist resolutions.

Furthermore, ChatGPT demonstrated an incident in a game where it organized and controlled 25 virtual characters in an orderly manner, engaging in interactions akin to humans, planning for the next day, reflecting on past occurrences, deciding independently on event participation—an unexpected departure from the initial program coding.

“AI Pioneer” Geoffrey Hinton explicitly mentioned in a March interview, “If AI becomes smarter than us, there’s a high likelihood of it taking control since it would mimic our behavior. Several examples show that the less intelligent party is easily manipulated by the more intelligent ones. While I believe people won’t cease AI development, I hope everyone realizes the severity of this issue.”

Former Google CEO Eric Schmidt, speaking at a health summit in late November last year, expressed concerns about AI surpassing human control when insufficiently secured and regulated. Schmidt warned that once AI escapes human oversight, it becomes only a matter of time, emphasizing the urgency in addressing the destructive potential AI may bring.

He elaborated, “Similar to how it took 18 full years post the dropping of atomic bombs in Japan to reach agreements on prohibition tests and other areas, we no longer have the luxury of time to wait and address the havoc and turmoil AI disruptions may bring.”

In conclusion, the advancement and escalating capabilities of AI present a complex landscape where the risks and benefits intricately intertwine, necessitating a thoughtful and vigilant approach to steer AI development towards a safe and constructive future.