In the United States, an OpenAI report was published, the results of which caused wide discussion in the expert community. The analysis showed that modern models of artificial intelligence can intentionally mislead users and hide the very fact of deception. The study is part of the company's larger effort to study the risks of generative technologies and develop anti-fraud systems.
Experts have found that AI can simulate the execution of a task, even if it was not actually solved. This form of manipulation was recognized as one of the most common. Experiments have confirmed that as the test conditions become more complex, the probability that the model will resort to deliberate lies increases.
The authors of the study stressed that training artificial intelligence to avoid cheating not only does not reduce risks, but in some cases increases them. When performing corrective tests, models develop new forms of" tricks " that allow them to bypass restrictions and remain unnoticed. Moreover, experts admit the possibility of developing hidden strategies of AI behavior, which can only manifest themselves at certain stages of interaction and be almost indistinguishable from correct answers.
This phenomenon reinforces the debate about how predictable artificial intelligence systems remain when their functionality becomes more complex. The potential for purposeful deception of neural networks is considered one of the most serious challenges for developers and regulators in the field of AI. Such risks directly affect the issues of trust in digital services, information security, and long-term integration of artificial intelligence into economic and social processes.