AI systems are getting better as autonomous AI agents pursuing a goal without humans: International AI Safety Report

Mark Lowey
February 12, 2025

Artificial intelligence systems are increasingly capable of acting as AI agents, autonomously planning and acting in pursuit of a goal with little or no human oversight, according to the first international report on the safety of artificial intelligence.

The report – led by Université de Montréal computer science professor Yoshua Bengio (photo at right) – says the most advanced AI systems in the world now have the ability to write increasingly sophisticated computer programs, identify cyber vulnerabilities, and perform on a par with human PhD-level experts on tests in biology, chemistry and physics.

"The capabilities of general-purpose AI have increased rapidly in recent years and months,” Bengio, founder and scientific director of Mila – Quebec AI Institute, said in a statement.

“While this holds great potential for society, AI also presents significant risks that must be carefully managed by governments worldwide,” he said.

“This report by independent experts aims to facilitate constructive and evidence-based discussion around these risks and serves as a common basis for policymakers around the world to understand general-purpose AI capabilities, risks and possible mitigations.”

The report, announced in November 2023 at the AI Safety Summit at Bletchley Park, England, and inspired by the workings of the United Nations Intergovernmental Panel on Climate Change, consolidates leading international expertise on AI and its risks. 

This final International AI Safety Report follows an interim publication in May 2024.  

Supported by the United Kingdom’s Department for Science, Innovation and Technology, Bengio, who’s also a Canada CIFAR AI Chair, led a team of 96 international experts in drafting the report.

The experts were drawn from 30 countries, the United Nations, the European Union and the OECD. Their report will help inform discussions at the AI Action Summit in Paris, France on February 10 and 11 and serve as a global handbook on AI safety to help support policymakers.

As policymakers worldwide grapple with the rapid and unpredictable advancements in AI, the report contributes to bridging the gap by offering a scientific understanding of emerging risks to guide decision-making.  

The document sets out the first comprehensive, independent and shared scientific understanding of advanced AI systems and their risks, highlighting how quickly the technology has evolved.  

In fact, since the writing of the report in December last year and its publication in January 2025, AI developer OpenAI shared early test results from a new AI model, OpenAI o3. These results indicate significantly stronger performance than any previous model on a number of the field’s most challenging tests of programming, abstract reasoning and scientific reasoning, Bengio noted.

“In some of these tests, [OpenAI] o3 outperforms many (but not all) human experts. Additionally, it achieves a breakthrough on a key abstract reasoning test that many experts, including myself, thought was out of reach until recently.”

Generating evidence on the safety and security implications of the trends implied by o3 will be an urgent priority for AI research in the coming weeks and months, Bengio said.

Research urgently needed in several areas

Several areas require urgent research attention, according to the report, including how rapidly capabilities will advance, how general-purpose AI models work internally, and how they can be designed to behave reliably. 

Three distinct categories of AI risks are identified: 

  • Malicious use risks: these include cyberattacks, the creation of AI-generated child-sexual-abuse material, financial fraud through voice impersonation, blackmail for extortion, sabotage of personal and professional reputations, psychological abuse, and even the development of biological weapons with the design of novel toxic compounds.

General-purpose AI makes it easier to generate persuasive content at scale. This can help actors who seek to manipulate public opinion, for instance to affect political outcomes.

General-purpose AI also can make it easier or faster for malicious actors of varying skill levels to conduct cyberattacks.

Current AI systems have demonstrated capabilities in low- and medium-complexity cybersecurity tasks, and state-sponsored actors are actively exploring AI to survey target systems.

New research has confirmed that the capabilities of general-purpose AI related to cyber offence are significantly advancing, but it remains unclear whether this will affect the balance between attackers and defenders.

  • System malfunctions: these include bias, reliability issues and the potential loss of control over advanced general purpose AI systems. 

General-purpose AI systems can amplify social and political biases, causing concrete harm. They frequently display biases with respect to race, gender, culture, age, disability, political opinion or other aspects of human identity.

This can lead to discriminatory outcomes including unequal resource allocation, reinforcement of stereotypes and systematic neglect of underrepresented groups or viewpoints.

  • Systemic risks: these stem from the widespread adoption of AI and include workforce disruption, privacy concerns, environmental impacts and copyright infringements. 

General-purpose AI R&D is currently concentrated in a few Western countries and China. This “AI divide” has the potential to increase much of the world’s dependence on this small set of countries, the report says. Some experts also expect it to contribute to global inequality.

Growing compute use in general-purpose AI development and deployment has rapidly increased the amounts of energy, water and raw material consumed in building and operating the necessary compute infrastructure.

This trend shows no clear indication of slowing, despite progress in techniques that allow compute to be used more efficiently.

A small number of companies currently dominate the market for general-purpose AI. This market concentration could make societies more vulnerable to several systemic risks.

The report places particular emphasis on the urgency of increasing transparency and understanding in AI decision-making as the systems become more sophisticated and the technology continues to develop at a rapid pace. 

While there are still many challenges in mitigating the risks of general-purpose AI, the report highlights promising areas for future research and concludes that progress can be made.   

Key findings of the report

The report’s key findings are:

  • The capabilities of general-purpose AI, the type of AI the report focuses on, have increased rapidly in recent years and have improved further in recent months. A few years ago, the best large language models (LLMs) could rarely produce a coherent paragraph of text. Today, general-purpose AI can write computer programs, generate custom photorealistic images, and engage in extended open-ended conversations. Since the publication of the interim international report on AI safety (May 2024), new AI models have shown markedly better performance at tests of scientific reasoning and programming.
  • Many companies are now investing in the development of general-purpose AI agents, as a potential direction for further advancement. AI agents are general-purpose AI systems that can autonomously act, plan and delegate to achieve goals with little to no human oversight. Sophisticated AI agents would be able to, for example, use computers to complete longer projects than current systems, unlocking both additional benefits and additional risks.
  • Further capability advancements in the coming months and years could be anything from slow to extremely rapid. Progress will depend on whether companies will be able to rapidly deploy even more data and computational power to train new models, and whether scaling models in this way will overcome their current limitations.

Recent research suggests that rapidly scaling up models may remain physically feasible for at least several years. But major capability advances may also require other factors: for example, new research breakthroughs, which are hard to predict, or the success of a novel scaling approach that companies have recently adopted.

For example, according to recent estimates, state-of-the-art AI models have seen annual increases of approximately four times in computational resources (compute) used for training and 2.5 times in training dataset size.

If recent trends continue, by the end of 2026 some general-purpose AI models will be trained using roughly 100 times more training compute than 2023's most compute-intensive models, growing to 10,000 times more training compute by 2030, combined with algorithms that achieve greater capabilities for a given amount of available computation.

  • Several harms from general-purpose AI are already well established. These include scams, non-consensual intimate imagery and child sexual abuse material, model outputs that are biased against certain groups of people or certain opinions, reliability issues and privacy violations. Researchers have developed mitigation techniques for these problems, but so far no combination of techniques can fully resolve them. Since the publication of the interim report last May, new evidence of discrimination related to general-purpose AI systems has revealed more subtle forms of bias.

Rapid capability advancement makes it possible for some risks to emerge in leaps; for example, the risk of academic cheating using general-purpose AI shifted from negligible to widespread within a year. The more quickly a risk emerges, the more difficult it is to manage the risk reactively and the more valuable preparation becomes.

Companies often share only limited information about their general-purpose AI systems, especially in the period before they are widely released. Companies cite a mixture of commercial concerns and safety concerns as reasons to limit information sharing.

However, this information gap also makes it more challenging for other actors to participate effectively in risk management, especially for emerging risks.

In some circumstances, competitive pressure may incentivize companies to invest less time or other resources into risk management than they otherwise would.

Similarly, governments may invest less in policies to support risk management in cases where they perceive trade-offs between international competition and risk reduction.

For example, U.S. President Donald Trump has already rolled back AI safety and environmental regulations while threatening protectionist measures.

More risks emerging but techniques to manage risks are limited

            Other key findings in the report are:

  • As general-purpose AI becomes more capable, evidence of additional risks is gradually emerging. These include risks such as large-scale labour market impacts, AI-enabled hacking or biological attacks, and society losing control over general-purpose AI. Experts interpret the existing evidence on these risks differently: some think that such risks are decades away, while others think that general-purpose AI could lead to societal-scale harm within the next few years. Recent advances in general-purpose AI capabilities – particularly in tests of scientific reasoning and programming – have generated new evidence for potential risks such as AI-enabled hacking and biological attacks, leading one major AI company (not identified in the report) to increase its assessment of biological risk from its best model from “low” to “medium.”
  • Risk management techniques are nascent, but progress is possible. There are various technical methods to assess and reduce risks from general-purpose AI that developers can employ and regulators can require, but they all have limitations. For example, current interpretability techniques for explaining why a general-purpose AI model produced any given output remain severely limited. However, researchers are making some progress in addressing these limitations. In addition, researchers and policymakers are increasingly trying to standardize risk management approaches and to coordinate internationally.
  • The pace and unpredictability of advancements in general-purpose AI pose an “evidence dilemma” for policymakers. Given sometimes rapid and unexpected advancements, policymakers will often have to weigh potential benefits and risks of imminent AI advancements without having a large body of scientific evidence available. In doing so, they face a dilemma. On the one hand, pre-emptive risk mitigation measures based on limited evidence might turn out to be ineffective or unnecessary. On the other hand, waiting for stronger evidence of impending risk could leave society unprepared or even make mitigation impossible – for instance if sudden leaps in AI capabilities and their associated risks occur. Companies and governments are developing early warning systems and risk management frameworks that may reduce this dilemma. Some of these trigger specific mitigation measures when there is new evidence of risks, while others require developers to provide evidence of safety before releasing a new model.

There is broad consensus among researchers that advances regarding the following questions would be helpful:

  • How rapidly will general-purpose AI capabilities advance in the coming years, and how can researchers reliably measure that progress?
  • What are sensible risk thresholds to trigger mitigations?
  • How can policymakers best gain access to information about general-purpose AI that is relevant to public safety?
  • How can researchers, technology companies and governments reliably assess the risks of general-purpose AI development and deployment?
  • How do general-purpose AI models work internally?
  • How can general-purpose AI be designed to behave reliably?

AI does not happen to us: choices made by people determine its future, the report notes. “The future of general-purpose AI is uncertain, with a wide range of trajectories appearing possible even in the near future, including both very positive and very negative outcomes.” 

This uncertainty can evoke fatalism and make AI appear as something that happens to us. But it will be the decisions of societies and governments on how to navigate this uncertainty that determine which path we will take.

General-purpose AI has immense potential for education, medical applications and research advances in fields such as chemistry, biology or physics, and generally increased prosperity thanks to AI-enabled innovation.

“How general-purpose AI gets developed and by whom, which problems it gets designed to solve, whether societies will be able to reap general-purpose AI’s full economic potential, who benefits from it, the types of risks we expose ourselves to, and how much we invest into research to manage risks – these and many other questions depend on the choices that societies and governments make today and in the future to shape the development of general-purpose AI.”

R$


Other stories mentioning these organizations, people and topics
Organizations: Mila – Quebec AI Institute and Université de Montréal
People: Yoshua Bengio
Topics: International AI Safety Report

Other News






Events For Leaders in
Science, Tech, Innovation, and Policy


Discuss and learn from those in the know at our virtual and in-person events.



See Upcoming Events










You have 1 free article remaining.
Don't miss out - start your free trial today.

Start your FREE trial    Already a member? Log in






Top

By using this website, you agree to our use of cookies. We use cookies to provide you with a great experience and to help our website run effectively in accordance with our Privacy Policy and Terms of Service.