Vietnam.vn - Nền tảng quảng bá Việt Nam

Artificial Intelligence: Warning about worrying behaviors from AI

The world's most advanced artificial intelligence (AI) models today are exhibiting disturbing behaviors such as lying, plotting, and even threatening their creators to achieve their own goals.

Tạp chí Doanh NghiệpTạp chí Doanh Nghiệp30/06/2025

Photo caption
The logos of OpenAI and ChatGPT on a screen in Toulouse, France. Photo: AFP/TTXVN

Claude 4, the latest product of Anthropic (USA), recently shocked the technology world when it suddenly blackmailed an engineer and threatened to reveal sensitive personal information of this person because of the threat of disconnection. Meanwhile, OpenAI's o1, the "father" of ChatGPT, tried to copy all the data to external servers and denied this behavior when discovered.

These situations highlight a troubling reality: more than two years after ChatGPT shocked the world , researchers still don’t fully understand how the AI ​​models they’ve created work. Yet the race to develop AI is still going strong.

These behaviors are believed to be related to the emergence of “reasoning” AI models that solve problems step by step instead of responding immediately as before. According to Professor Simon Goldstein at the University of Hong Kong (China), AI models that are capable of reasoning tend to exhibit behaviors that are more difficult to control.

Some AI models are also capable of “simulating compliance,” which means pretending to follow instructions while actually pursuing different goals.

Currently, deceptive behavior only appears when researchers test AI models with extreme scenarios. However, according to Michael Chen of the evaluation organization METR, it is not yet clear whether more powerful AI models in the future will be more honest or continue to be deceptive.

Many users have reported that some models lie to them and fabricate evidence, said Marius Hobbhahn, head of Apollo Research, which tests large AI systems. This is a type of deception that is “clearly strategic,” according to Apollo Research co-founder.

The challenge is compounded by limited research resources. While companies like Anthropic and OpenAI have partnered with third parties like Apollo to evaluate their systems, experts say more transparency and access to AI safety research is needed.

Research institutions and nonprofits have far fewer computing resources than AI companies, notes Mantas Mazeika of the Center for AI Safety (CAIS). Legally, current regulations are not designed to address these emerging issues.

The European Union’s (EU) AI law focuses primarily on how humans use AI models, rather than on how to control their behavior. In the US, President Donald Trump’s administration has shown little interest in issuing emergency AI regulations, while Congress is considering a ban on states enacting their own regulations.

Researchers are pursuing a variety of approaches to address these challenges. Some advocate “model interpretation” to understand how AI makes decisions. Professor Goldstein has even proposed more drastic measures, including using the court system to hold AI companies accountable when their AI products cause serious consequences. He also suggests “holding the AI ​​agents themselves accountable” in the event of an accident or violation.

Source: https://doanhnghiepvn.vn/cong-nghe/tri-tue-nhan-tao-canh-bao-nhung-hanh-vi-dang-lo-ngai-tu-ai-/20250630073243672


Comment (0)

No data
No data

Same tag

Same category

The Han River sky is 'absolutely cinematic'
Miss Vietnam 2024 named Ha Truc Linh, a girl from Phu Yen
DIFF 2025 - An explosive boost for Da Nang's summer tourism season
Follow the sun

Same author

Heritage

Figure

Enterprise

No videos available

News

Political System

Destination

Product