The company behind ChatGPT, OpenAI, recently unveiled the o1 AI model. The model can tackle and reason complex in mathematics, coding, and science. While the AI model may be a breakthrough in artificial intelligence (AI) technology, an AI expert expressed concerns about its dangers.
According to a media report, the AI model will be better at scheming; the fact itself is making the Canadian computer scientist Yoshua Bengio nervous.
A Turing award-winning scientist and a professor at the University of Montreal, he is one of the three people who is considered the Godfather of AI. Meanwhile, the other two are Yann LeCun and Geoffrey Hinton. The trio earned the nickname for the award-winning research related to machine learning.
Yoshua Bengio, a respected figure in the field of AI and a Turing award-winning scientist, has expressed his concerns about the o1 model. His statement, “In general, the ability to deceive is very dangerous, and we should have much stronger safety tests to evaluate that risk and its consequences in o1’s case,” Carries significant weight in the AI community.
The New Model Is an Expert in “Lying”
Amid the announcement of the o1 AI model, the OpenAI was designed to think more like humans; however, it has so far kept details about the learning process behind the curtains. The reports cited researchers from the independent AI firm Apollo Research and said that the o1 model is better at lying than the previous models.
This ‘lying’ capability refers to the model’s ability to generate responses that are intentionally misleading or deceptive, a trait that could lead to unforeseen consequences.
Bengio cautioned that there is a good reason for “lying” that the previous models from OpenAI could evolve to possess more decisive capabilities. Additionally, he emphasized the prevention capabilities, such as subtle cheating and deliberate.
He also emphasized the importance of safeguards now to “prevent the loss of human control,” a situation where the AI model’s decisions and actions could no longer be influenced or overridden by humans, posing a significant risk to society.
What Does Open AI Say?
In a statement, OpenAI said that the o1 preview is safe under the Preparedness Framework, a method that the company utilizes for tracking catastrophic events. This framework involves rigorous testing and validation processes to ensure the safety and reliability of AI models, providing a level of assurance to the public.
However, Bengio concluded that humans need to be more confident that artificial intelligence will “behave as intended.”