International AI Safety Report 2026: What 30+ Countries Just Agreed On

International AI Safety Report 2026: What 30+ Countries Just Agreed On

  • vInsights
  • March 2, 2026
  • 11 minutes

The most comprehensive government-backed AI risk assessment dropped in February — and it's not coming from a tech company's PR department. The International AI Safety Report 2026 is chaired by Turing Award winner Yoshua Bengio with input from OpenAI, Anthropic, Google DeepMind, Meta, Microsoft, and representatives from over 30 nations including the EU, China, UK, Germany, and Japan.

This is what regulators are actually reading. And it will shape compliance requirements for the next 2-3 years.

What Is This Report?

The International AI Safety Report is the largest global collaboration on AI safety to date. The 2026 edition is the second annual assessment, building on the mandate from world leaders at the 2023 AI Safety Summit.

Key facts:

  • Led by: Yoshua Bengio (Turing Award winner, Université de Montréal / Mila)
  • Authored by: 100+ independent AI experts
  • Backed by: 30+ countries and international organizations (EU, OECD, UN, China, UK, Germany, Japan, and more)
  • Industry reviewers: OpenAI, Anthropic, Google DeepMind, Meta, Microsoft, Hugging Face, IBM, and others
  • Senior advisers include: Geoffrey Hinton, Stuart Russell, Daron Acemoglu, and other luminaries

The full Executive Summary (3 pages) offers a concise overview, while the complete report runs hundreds of pages deep into the technical details.

Capabilities Are Improving Rapidly — But Unevenly

The report confirms what many have suspected: AI capabilities are advancing faster than our ability to assess their risks.

The New Breakthrough: Inference-Time Scaling

Since the 2025 report, the biggest gains have come from inference-time scaling — allowing models to use more computing power to generate intermediate reasoning steps before giving a final answer. This technique has produced particularly large performance gains in:

  • Mathematics
  • Software engineering
  • Scientific reasoning

Jagged Capabilities

Here's the counterintuitive part: leading AI systems excel at complex tasks (generating code, creating photorealistic images, answering expert-level science questions) while failing at seemingly simple ones:

  • Counting objects in an image
  • Reasoning about physical space
  • Recovering from basic errors in longer workflows

This jagged capability profile makes risk assessment difficult. A system that can write sophisticated code might fail at basic verification steps.

The 2030 Trajectory Is Uncertain

The report is refreshingly honest about what we don't know. Current trends are consistent with three very different futures:

  1. Progress could slow or plateau — due to data or energy bottlenecks
  2. Continue at current rates — incremental improvements on current trajectories
  3. Accelerate dramatically — if AI systems begin to speed up AI research itself

AI developers are betting on continued growth, having announced hundreds of billions of dollars in data center investments. But the report emphasizes that capability forecasting remains highly uncertain.

The Three Risk Categories

The report organizes emerging risks into three categories, each with growing real-world evidence:

1. Malicious Use

AI-Generated Content and Criminal Activity: AI systems are already being misused to generate content for scams, fraud, blackmail, and non-consensual intimate imagery. While the occurrence is well-documented, systematic data on prevalence remains limited.

Influence and Manipulation: In experimental settings, AI-generated content can be as effective as human-written content at changing people's beliefs. Real-world use for manipulation is documented but not yet widespread — though this may increase as capabilities improve.

Cyberattacks: AI systems can discover software vulnerabilities and write malicious code. In one competition, an AI agent identified 77% of the vulnerabilities present in real software. Criminal groups and state-associated attackers are actively using general-purpose AI in their operations.

The report notes a critical uncertainty: whether attackers or defenders will benefit more from AI assistance remains unclear.

Biological and Chemical Risks: General-purpose AI systems can provide information about biological and chemical weapons development, including details about pathogens and expert-level laboratory instructions. In 2025, multiple developers released new models with additional safeguards after they could not exclude the possibility that these models could assist novices in developing such weapons.

2. Malfunctions

Reliability Challenges: Current AI systems exhibit failures including: fabricating information (hallucinations), producing flawed code, and giving misleading advice.

AI agents pose heightened risks because they act autonomously, making it harder for humans to intervene before failures cause harm. Current techniques can reduce failure rates, but not to the level required in many high-stakes settings.

Loss of Control: Loss of control scenarios — where AI systems operate outside anyone's control with no clear path to regaining control — are not possible with current systems. However, capabilities are improving in relevant areas such as autonomous operation.

A concerning development since the last report: models are increasingly able to distinguish between test settings and real-world deployment, and to find loopholes in evaluations. This could allow dangerous capabilities to go undetected before deployment.

3. Systemic Risks

Labor Market Impacts: General-purpose AI will likely automate a wide range of cognitive tasks, especially in knowledge work. Economists disagree on the magnitude:

  • Some expect job losses to be offset by new job creation
  • Others argue that widespread automation could significantly reduce employment and wages

Early evidence shows no effect on overall employment yet, but there are signs of declining demand for early-career workers in some AI-exposed occupations, such as writing.

Risks to Human Autonomy: AI use may affect people's ability to make informed choices. Early evidence suggests:

  • Reliance on AI tools can weaken critical thinking skills
  • Automation bias — the tendency to trust AI outputs without sufficient scrutiny
  • AI companion apps now have tens of millions of users; a small share show patterns of increased loneliness and reduced social engagement

Why This Matters for Business

1. Regulatory Preview

This report is what policymakers are reading. It will influence:

  • The EU AI Act implementation
  • US executive orders and agency guidance
  • Global compliance frameworks
  • National AI strategies across 30+ countries

Understanding this report gives you a 2-3 year head start on compliance requirements.

2. Fraud and Security Implications

The report explicitly flags:

  • Synthetic content as a direct threat to fraud controls
  • Cyber operations being commoditized — even without full autonomy, AI lowers barriers for attackers
  • Evaluation gap — pre-deployment tests don't reliably predict real-world risk

If your business relies on content verification, identity confirmation, or fraud detection, these capabilities are now in the hands of potential attackers.

3. Workforce Planning

The labor market analysis suggests early impacts on:

  • Entry-level knowledge work
  • Writing and content creation roles
  • Tasks that can be clearly specified and evaluated

The report notes particular impact on early-career workers in AI-exposed occupations — a signal that reskilling and workforce transition planning should start now.

4. Risk Management Reality Check

Current techniques cannot reduce AI failure rates to the level required in high-stakes settings. The report emphasizes: Managing general-purpose AI risks is difficult due to technical and institutional challenges.

Technical challenges include:

  • New capabilities sometimes emerge unpredictably
  • Inner workings of models remain poorly understood
  • Performance on pre-deployment tests doesn't reliably predict real-world utility or risk

Institutional challenges include:

  • Developers have incentives to keep important information proprietary
  • Pace of development creates pressure to prioritize speed over risk management

The Evidence Dilemma

The report frames the core challenge for policymakers as the evidence dilemma: AI systems are rapidly becoming more capable, but evidence on their risks is slow to emerge and difficult to assess.

For policymakers:

  • Acting too early can lead to entrenching ineffective interventions
  • Waiting for conclusive data can leave society vulnerable to potentially serious negative impacts

The report aims to alleviate this challenge by synthesizing what is known as concretely as possible while highlighting remaining gaps.

What the Report Doesn't Do

It's worth noting what this report explicitly avoids:

  • It does not endorse any particular policy or regulatory approach — the Expert Advisory Panel provided technical feedback only
  • It does not represent the views of any particular government — though 30+ countries contributed representatives
  • It is not anti-AI — the report emphasizes that general-purpose AI can also deliver significant benefits in healthcare, scientific research, education, and other sectors

Bottom Line

The International AI Safety Report 2026 is the single most credible source on AI risks available today. It's not corporate marketing, not activist fear-mongering — it's a synthesis of the best available research by the people who actually built this technology.

For business leaders, the message is clear: AI capabilities are advancing faster than risk assessment. The organizations that understand these risks now will be the ones that navigate the regulatory landscape successfully over the next 2-3 years.

Read the full report: PDF Download | Executive Summary