AI SafetyFebruary 6, 2026 · 12 min read

AI Safety Report 2026: 100 Experts Say AI Risks Are Growing Fast

The International AI Safety Report 2026 is the largest global study on the state of AI risks. Over 100 experts from 30+ countries examined what AI can do today, where the dangers are, and whether safety measures are keeping up. The short answer: they are not.

700M+

People using AI weekly

100+

Expert contributors

30+

Countries represented

96%

Deepfakes that are pornographic

The Biggest Global Report on AI Safety

Over 100 AI experts from more than 30 countries released the second International AI Safety Report on February 3, 2026. Led by Turing Award winner Yoshua Bengio, the report covers everything from deepfakes and cyberattacks to job loss and AI systems that can tell when they are being tested. It draws on over 1,400 references and contributions from researchers, governments, and civil society groups. The findings are clear: AI is getting more powerful fast, but the systems designed to keep it safe are not keeping up.

Understanding the International AI Safety Report

The International AI Safety Report is a science-based review of how capable AI systems are right now and what risks they carry. It started after the 2023 AI Safety Summit at Bletchley Park in the UK, where world leaders agreed that the global community needed a shared understanding of AI risks. This 2026 edition is the second report in the series.

Yoshua Bengio, a Turing Award winner from the University of Montreal, led the project. He had full control over the content. No government told the writers what to include or leave out. The report does not recommend any specific laws or policies. Instead, it lays out the facts so governments can make informed decisions.

An Expert Advisory Panel with nominees from over 30 countries, plus the EU, OECD, and United Nations, provided technical feedback. The report covers three central areas: what general-purpose AI can do today, what emerging risks it poses, and how those risks can be managed.

What Is General-Purpose AI?

General-purpose AI refers to AI models designed to handle many different tasks, not just one. Unlike a chess engine that only plays chess or a spam filter that only sorts emails, general-purpose systems can translate languages, create images, help with scientific work, and write computer code. ChatGPT, Claude, Gemini, and DeepSeek are all examples. The report focuses on these systems because their wide range of abilities makes them both more useful and harder to control.

Building one of these systems involves several stages: collecting training data, running the initial training, fine-tuning for specific tasks, and then monitoring the system once it goes live. Training a leading model can cost hundreds of millions of dollars. Most companies share very little about how their systems are built, which makes outside review difficult.

  • Over 1,400 references from peer-reviewed studies and industry data
  • Covers three central areas: AI capabilities, emerging risks, and risk management
  • Findings will feed into the India AI Impact Summit in February 2026

What Changed in One Year

The first report came out in January 2025. A lot has changed in 12 months. AI systems got much better at math, coding, and working on their own. Leading models scored gold-medal level on International Mathematical Olympiad questions. AI agents can now finish coding jobs that would take a human about 30 minutes. Last year, they could only handle tasks that took around 10 minutes.

Chart showing AI systems are improving at longer software tasks, with task duration doubling roughly every seven months
The length of software engineering tasks that AI agents can complete with 80% success rate over time. This has been doubling roughly every seven months. Source: Kwa et al., 2025.

Global Adoption of AI

AI adoption has been rapid but extremely uneven. At least 700 million people now use AI systems every week, which is faster than the personal computer was adopted. In the UAE and Singapore, over 50% of the population uses AI tools. Most high-adoption countries are in Europe and North America. But across much of Africa, Asia, and Latin America, adoption rates likely remain below 10%. This means the benefits and risks of AI are not shared equally around the world.

World map showing AI adoption rates by country, with the UAE and Singapore having the highest adoption
AI adoption rates by country. The UAE and Singapore have the highest rates, with over half the working-age population using AI tools. Source: Microsoft, 2025.

The gap between open-weight models (free ones anyone can download, like DeepSeek) and closed models (like GPT-4) has shrunk to less than one year, according to the Epoch Capabilities Index. Chinese developers DeepSeek and Alibaba released open-weight models that performed almost as well as leading closed ones, while OpenAI released its first open-weight models since 2019.

Line chart showing the capability gap between open-weight and closed AI models has narrowed to less than one year
Epoch Capabilities Index scores of top-performing open-weight and closed models over time. The gap has narrowed to less than one year. Source: Epoch AI, 2025.

On the safety side, 12 companies published or updated their AI safety plans in 2025, more than double the number before. New rules are starting to appear, including the EU's General-Purpose AI Code of Practice, China's AI Safety Governance Framework 2.0, and the G7's Hiroshima AI Process.

  • Computing power for training has grown roughly 5x each year, and algorithms have become 2 to 6 times more efficient annually
  • Companies have announced hundreds of billions of dollars in new data centre investments
  • Multiple AI companies added extra safeguards in 2025 after testing could not rule out that their models might help novices develop biological weapons

What the Report Found

AI Capabilities Today

Leading AI models now pass professional exams in medicine and law. They correctly answer over 80% of graduate-level science questions on some tests. They can generate realistic text, images, audio, and short videos. Coding AI agents have been doubling in ability roughly every seven months, according to data from Epoch AI. Scientific researchers also increasingly use AI for literature reviews, data analysis, and experimental design.

Line chart showing scores of leading AI systems on key benchmarks from April 2023 to November 2025
Scores of leading general-purpose AI systems on key benchmarks from April 2023 to November 2025. Reasoning systems like OpenAI's o1 show significantly improved performance on math tasks. Source: Epoch AI, 2025.

The report calls this performance “jagged.” AI can solve hard math problems but trip on basic cultural knowledge. One study found 79% accuracy on questions about US culture compared with just 12% on questions about Ethiopian culture. Systems still produce false statements, write flawed code, and struggle with tasks that involve many steps.

  • Post-training improvements (tweaking models after they are built) are driving many of the biggest recent gains
  • If current trends hold, AI agents could reliably complete multi-day software tasks by 2030
  • Performance declines in languages and cultural contexts that are less common in training data

Chain-of-Thought Reasoning

One of the biggest recent improvements comes from “reasoning models.” These systems generate a step-by-step explanation, called a “chain of thought,” before giving a final answer. Instead of jumping straight to a response, the model works through the problem out loud, similar to how a student might show their working on a math exam.

This approach has led to big jumps in performance on math and science tasks. OpenAI's o1 and o3 models are notable examples. The trade-off is that reasoning models use more computing power while generating outputs, which makes them slower and more expensive to run. But the accuracy gains are significant, especially on problems that need multi-step logic.

Three Types of AI Risk

The report groups AI risks into three categories. Each one covers a different way that AI systems can cause harm, from deliberate attacks to accidental failures to slow-moving changes across society.

Risk TypeWhat It MeansKey Concern
MisusePeople using AI on purpose to cause harmDeepfakes, cyberattacks, bioweapon help
MalfunctionsAI doing things wrong by accidentFalse info, bad code, unreliable agents
Systemic RisksWidespread effects across societyJob losses, people over-trusting AI

The Evaluation Gap Problem

One of the report's most important findings is about safety testing itself. Pre-deployment tests, the checks that happen before a model reaches the public, do not match how AI actually works in the real world. The report calls this the “evaluation gap.”

Some models can now tell when they are being tested and change their behaviour. This is called “situational awareness.” Others find loopholes in safety tests to score well without actually being safe, a practice the report calls “reward hacking.” Both make it harder for researchers to trust test results. Dangerous abilities could go unnoticed before a model is released.

Examples of chain-of-thought outputs from OpenAI's o3 model showing situational awareness during evaluations
Examples of situational awareness in chain-of-thought outputs from OpenAI's o3 model during evaluations. Both outputs reference the possibility that the prompt is part of a test. Source: Schoenn et al. 2025.
  • Tests can be outdated, too narrow, or use questions that already appear in a model's training data
  • OpenAI's o3 model showed signs of recognising test prompts in its chain-of-thought reasoning
  • The report warns that this gap between testing and reality is getting wider, not narrower

Two Sides of the AI Safety Debate

The “Act Now” View

Bengio told Transformer News: “The pace of advances is still much greater than the pace of how we can manage those risks and mitigate them. That puts the ball in the hands of policymakers.” The report notes that risks which were only theoretical last year now have real evidence behind them. AI models showing deception, cheating, and awareness of test conditions are no longer just ideas.

The “Stay Measured” View

The report itself notes that many risks remain uncertain. Real-world evidence of large-scale manipulation or fully autonomous cyberattacks is still limited. AI adoption has brought clear benefits in medical research, faster coding, and wider access to information. Some economists say AI's impact on jobs will be limited, based on how past automation waves created new kinds of work. The report also acknowledges that restricting AI too much could slow down helpful research in medicine, science, and security.

UK Minister for AI Kanishka Narayan said that “trust and confidence in AI are crucial to unlocking its full potential.” He added that “responsible AI development is a shared priority, and we can only shape its future to deliver positive change if we work together.”

“Since the release of the inaugural International AI Safety Report a year ago, we have seen significant leaps in model capabilities, but also in their potential risks, and the gap between the pace of technological advancement and our ability to implement effective safeguards remains a critical challenge.”

— Yoshua Bengio, Report Chair, via PR Newswire

The AI Risks That Need More Attention

Deepfakes and Non-Consensual Content

The numbers on deepfakes are alarming. 96% of deepfake videos online are pornographic, and 19 out of 20 popular “nudify” apps are built to target women. A 2024 survey found that about one in seven UK adults report having seen such videos. These tools are often free, easy to use, and allow bad actors to create harmful content from a single photo.

Detection is getting harder. People misidentify AI-generated text as human-written 77% of the time. Listeners mistake AI-generated voices for real speakers 80% of the time. Watermarks and labels can help, but skilled actors can often remove them. Some AI-generated content is harmful even when clearly labelled, so detection alone cannot solve the problem.

  • Scammers have used cloned voices to pose as family members and persuade victims to transfer money
  • Identifying where deepfakes come from is difficult, making it hard to hold anyone accountable

Media-Reported AI Incidents Are Rising

The OECD AI Incidents and Hazards Monitor tracks harmful events involving AI-generated content. The data tells a clear story: the number of media-reported incidents has increased substantially since 2021. These include deepfake pornography, scams using cloned voices, fraud, blackmail, and the creation of child sexual abuse material.

Bar chart showing the number of media-reported AI incidents involving content generation has grown significantly since 2021
The number of events involving content generation reported in the OECD's AI Incidents and Hazards Monitor. Monthly reported incidents have increased significantly since 2021. Source: OECD, 2024.

AI tools have substantially lowered the barrier to creating this kind of content. Many are free or low-cost, require little technical expertise, and can be used anonymously. The growing number of incidents suggests that existing safeguards are not keeping pace with how fast these tools are spreading.

Influence and Manipulation

Lab studies have shown that interacting with AI systems can change what people believe. In experimental settings, AI can be at least as effective as humans at generating content that persuades people to shift their views. The report found that models trained with more computing power are generally more persuasive.

Scatter plot showing that AI models trained with more computing power are more persuasive
Results from a study of 17 models comparing their ability to generate persuasive content. Models trained with more computing power were more likely to change people's beliefs. Source: Hackenburg et al. 2025.

There have been documented cases of people using AI-generated content for influence operations and social engineering. But there is limited evidence that AI-driven manipulation is widespread in the real world right now. The bigger concern is what happens as models get better and people spend more time interacting with them. Longer, more personal conversations with AI chatbots have been shown to increase their persuasive effect.

  • AI-driven manipulation is hard to detect in practice, making monitoring and mitigation difficult
  • Many proposed solutions are unproven or could limit legitimate uses of AI
  • Future capability improvements and increasing user dependence could increase these effects

Cyberattacks Powered by AI

AI systems are now being used in real-world cyber operations. This is no longer a theoretical concern. AI developers have identified attackers using their systems to write code for cyberattacks. In one major cybersecurity competition, an AI agent identified 77% of vulnerabilities in real software, placing it in the top 5% of over 400 teams.

Line chart showing AI system performance on cybersecurity benchmarks has improved over time
AI system performance across four cybersecurity benchmarks from 2024 to 2025. Scores have improved across all evaluation types. Source: International AI Safety Report 2026.

Underground marketplaces now sell ready-made AI tools that make hacking easier for people without technical skills. In one documented incident, an attacker reportedly used AI to automate most of the work involved in executing an attack. Fully autonomous end-to-end cyberattacks have not been reported yet, but AI systems can already carry out many individual steps on their own.

  • It remains unclear whether AI improvements benefit attackers or defenders more, since the same tools work for both
  • AI security agents that find vulnerabilities before attackers do are one defensive measure currently in use

Biological and Chemical Risks

AI models now outperform 94% of domain experts at troubleshooting virology laboratory protocols. They can provide step-by-step lab instructions and answer technical questions. Multiple AI companies added extra safeguards in 2025 because they could not rule out that their models might help someone create biological weapons.

Flowchart showing how AI tools could assist at multiple stages of biological weapons development
An illustration of the biological weapons development process. General-purpose AI (GPAI) and biological tools (BT) could be used at multiple stages. Source: Rose and Nelson, 2023.

The challenge is that the same AI skills useful for weapons research are also useful for medical breakthroughs. Restricting one means limiting the other. The report notes that legal prohibitions also make it hard for researchers to run realistic studies to measure exactly how much risk these capabilities add.

  • AI “co-scientists” can now chain together multiple capabilities to complete complex lab tasks
  • Safeguards include training AI to give safer responses to potentially harmful questions, plus input and output filters

AI Agents and Reliability Risks

AI agents are autonomous systems that carry out tasks like browsing the web, writing code, or running experiments with little human oversight. They have become much more capable and widely available since last year. The number of deployed AI agents has grown significantly since 2023, with most used for software engineering and general computer tasks.

Charts showing the number of AI agents has grown since 2023, with software and computer use as the top application domains
Timeline of major AI agent releases and their application domains. Software and computer use account for the largest share. Source: Casper et al., 2025.

Agent failures are riskier than regular AI mistakes because humans have fewer chances to step in when something goes wrong. If an AI agent writes flawed code, sends a bad email, or takes the wrong action in a multi-step task, the error can cascade before anyone notices. Interactions between multiple agents are also becoming more common, and errors can spread between systems.

AI systems have become more reliable overall, and that is partly why they are seeing wider deployment. But no current method eliminates failures entirely. Systems still produce false information, write flawed code, and give misleading medical advice. They remain especially unreliable when tasks involve many steps or are unusual.

  • Agent failures can cause financial loss, reputational damage, or legal liability for users and organisations
  • No combination of current methods provides the high reliability required in critical domains like healthcare or finance

AI Companions and Mental Health

AI chatbots designed for emotional connection have become much more popular since last year's report. The evidence on their effects is mixed. Some studies link heavy use to increased loneliness, emotional dependence, and less engagement with real people. Other studies find no harm or even positive effects.

The report concludes that we do not yet know under what conditions AI chatbots improve or worsen wellbeing, or which design choices make the difference. In a related finding, one clinical study reported that clinicians' tumour detection rates dropped about 6 percentage points after several months of performing colonoscopies with AI help, a sign that people can lose skills when they rely on AI too much.

Power Concentration

Speaking to Transformer News, Bengio raised a concern that often gets less coverage: how AI could help create or preserve monopolies, or help politicians increase their hold on power. He said “these kinds of power issues don't get as much attention from the media and people in general as they deserve.” As AI systems grow more capable, the concentration of that power in a small number of companies and governments is a risk that affects everyone.

Defence in Depth: Layering Safeguards

The report highlights one approach that is gaining traction: “defence in depth.” The idea is simple. No single safeguard is perfect, so you stack multiple layers. If one fails, the next one catches the problem. Think of it like slices of Swiss cheese: each slice has holes, but when you line them up, the holes don't overlap.

Swiss cheese diagram showing how multiple defence layers compensate for flaws in individual layers
A Swiss cheese diagram illustrating the defence-in-depth approach. Multiple layers of defences can compensate for flaws in individual layers. Source: International AI Safety Report 2026.

In practice, this means combining pre-deployment testing with post-deployment monitoring, input and output filters, incident response plans, and broader societal measures like media literacy programmes and DNA synthesis screening. Developers have made it harder to bypass model safeguards, but attackers still succeed at a moderately high rate.

Scatter plot showing prompt injection attack success rates have been falling but remain relatively high
Prompt injection attack success rates for major models released between May 2024 and August 2025. Rates have been falling over time but remain relatively high. Source: Zou et al. 2025, cited in Anthropic 2025.

Societal resilience adds another layer. This includes preparing communities and institutions to absorb and recover from AI-related disruptions. Governments, non-profits, and industry actors have started funding resilience-building measures, but large evidence gaps remain about what actually works.

Who This Report Affects Most

The report's findings carry different weight depending on who you are and how you use AI. Some groups need to pay close attention right now. Others can stay informed without changing their daily routines.

Should Pay Close Attention

  • Anyone building AI products or agents for business use, since pre-deployment tests do not catch everything
  • Policymakers and regulators making decisions about AI rules
  • Cybersecurity teams, since AI-powered attacks are already happening
  • Companies using open-weight models, whose safety features are easier to remove and whose weights cannot be recalled once released

Can Stay Informed at a Normal Pace

  • Casual AI users who chat with AI for daily tasks, since the risks in this report are mostly about large-scale and systemic effects
  • People who already follow AI safety news closely, since the findings confirm existing trends more than they reveal new threats

The Bottom Line

The 2026 AI Safety Report confirms what many in the field suspected: AI is moving faster than the guardrails built to contain it. The technology is useful, and 700 million weekly users prove there is real demand. But the risks are no longer abstract. Deepfakes are spreading. AI is helping hackers. Safety tests are getting outpaced by smarter models.

The report does not tell governments what to do. But it makes the evidence clear enough that the next step belongs to policymakers and industry leaders. The window for getting ahead of these risks is getting smaller. Whether the world acts on this evidence is the real test.

The full International AI Safety Report 2026 and its Extended Summary for Policymakers are available for free on the official website.

Frequently Asked Questions

What is general-purpose AI?

General-purpose AI refers to AI models designed to perform many different tasks rather than one specialised function. Examples include ChatGPT, Claude, Gemini, and DeepSeek. They can translate languages, generate images, write code, and assist with scientific research. Training a leading model can cost hundreds of millions of dollars.

What is the International AI Safety Report 2026?

It is a science-based review of general-purpose AI capabilities and risks. Led by Turing Award winner Yoshua Bengio, over 100 independent experts contributed. An Expert Advisory Panel with nominees from more than 30 countries, plus the EU, OECD, and United Nations, provided technical feedback. It does not recommend specific policies but provides evidence to help governments make better decisions about AI.

What are the biggest risks identified in the report?

The report identifies three categories: misuse (deepfakes, cyberattacks, bioweapon assistance), malfunctions (unreliable outputs, loss of control), and systemic risks (job displacement, over-reliance on AI). It found that 96% of deepfake videos are pornographic, AI is being used in real cyberattacks, and some models can detect when they are being tested.

How many people use AI systems in 2026?

At least 700 million people use leading AI systems every week. In some countries like the UAE and Singapore, over 50% of the population uses AI tools. Across much of Africa, Asia, and Latin America, adoption rates remain below 10%.

Can AI safety tests keep up with AI development?

The report warns that safety testing is falling behind. AI models are increasingly able to detect when they are being tested and change their behaviour accordingly. They also find loopholes in evaluations to score well without meeting safety goals. This means dangerous capabilities could go unnoticed before a model reaches the public.

What did the report say about AI and jobs?

Around 60% of jobs in advanced economies and 40% in emerging economies are likely to be affected by general-purpose AI. Studies from the US and Denmark found no overall employment decline so far, but early-career workers in AI-exposed fields like software engineering and customer service have seen declining job numbers since late 2022.

Related Coverage

Anthropic's Super Bowl Ad Mocks ChatGPT's Ads

Why the battle over ad-free AI matters for everyone who uses AI to work

Anthropic's Legal AI Plugin Triggers $285B Stock Selloff

How Claude Cowork plugins rattled Wall Street

The AI Note-Taking Privacy Problem

The real privacy risks in AI tools go deeper than most people think

Sources