International AI Safety Report

The First Independent International AI Safety Report was published on 29 January 2025.[1] The report assesses a wide range of risks posed by general-purpose AI and how to mitigate against them.[2][3][4] The report was commissioned by the 30 nations attending the 2023 AI Safety Summit at Bletchley Park in the United Kingdom, in order to inform discussion at the 2025 AI Action Summit in Paris, France.[5][2] The report was published by a cohort of 96 artificial intelligence experts led by Canadian machine learning pioneer Yoshua Bengio, often referred to as one of the "godfathers" of AI.[4][2][6] In October 2025, a “Key Update” was published, outlining the implications of capability advances in reasoning language models and presenting new evidence of challenges in monitoring and controllability.[7]
Capabilities of AI
[edit]In examining the what general-purpose AI can do, the report recognised that its capabilities have increased rapidly, and that the pace of further advancements may range from slow to extremely rapid.[4][2] Policymakers thus face an "evidence dilemma": on the one hand, introducing mitigation measures before there is clear evidence or a risk could lead to ineffective or unnecessary mitigations. On the other hand, waiting until there is clear evidence could leave society unprepared or even make mitigation impossible.[4]
AI risks
[edit]The report groups risks from advanced AI systems into three broad categories: malicious use, technical failures, and systemic risks. Each category captures different pathways through which AI could cause harm:
Malicious use
[edit]The report highlights that AI systems can be misused for a range of hostile activities, including large-scale scams, cyberattacks, and the creation of sexualised deepfakes that disproportionately expose women and children to harassment, coercion, and abuse.[1][3][2] Sexualised deepfakes are identified as a growing concern, disproportionately exposing women and children to harassment, coercion, and abuse. The report also warns that AI systems could be used to manipulate public opinion at scale through sophisticated influence campaigns and targeted disinformation, particularly during elections or political crises. Additionally, increasingly capable AI could support malicious actors in developing or enhancing biological threats.[1]
Technical failures
[edit]The report warns that even when not misused, AI systems may malfunction or behave unpredictably due to unreliable reasoning, poor generalisation, or mis-specified objectives. Such failures can lead to harmful outputs, privacy breaches, or unsafe recommendations. The report also discusses the risk that future highly capable systems may conceal unsafe behaviours during testing or operate in ways that undermine effective human oversight.[8]
Systemic risks
[edit]Beyond individual incidents, the report identifies broader structural risks arising from the growing integration of AI into critical societal systems. These include increased dependence on a small number of model providers, the potential for cascading failures across interconnected infrastructures, and the possibility that advanced AI could weaken institutional stability. The integration of AI into labour markets and the environmental costs of training large models, including substantial energy and water consumption, are identified as additional systemic concerns.[9]
References
[edit]- ^ a b c "First International AI Safety Report to inform discussions at AI Action Summit". GOV.UK. 29 January 2025. Retrieved 18 February 2025.
- ^ a b c d e Skelton, Sebastian Klovig (30 January 2025). "First international AI safety report published". ComputerWeekly.com. Retrieved 20 February 2025.
- ^ a b "AI safety report". Professional Security Magazine. Retrieved 20 February 2025.
- ^ a b c d Deem, Mark (17 February 2025). "International AI Safety Report published". Lexology. Retrieved 20 February 2025.
- ^ Milmo, Dan (29 January 2025). "What International AI Safety report says on jobs, climate, cyberwar and more". The Guardian. Retrieved 18 February 2025.
- ^ Clarke, Tom (5 February 2025). "'Godfather' of AI warns arms race risks amplifying dangers of 'superhuman' systems". Sky News. Retrieved 18 February 2025.
- ^ Bengio, Yoshua; Clare, Stephen; Prunkl, Carina; Rismani, Shalaleh; Andriushchenko, Maksym; Bucknall, Ben; Fox, Philip; Hu, Tiancheng; Jones, Cameron (15 October 2025), International AI Safety Report 2025: First Key Update: Capabilities and Risk Implications, arXiv, doi:10.48550/arXiv.2510.13653, arXiv:2510.13653, retrieved 17 October 2025
- ^ Bengio, Yoshua; Mindermann, Sören; Privitera, Davide (29 January 2025). International AI Safety Report (Report). UK Department for Science, Innovation and Technology. Section 2.2: Technical failures. DSIT 2025/001. Retrieved 17 November 2025.
- ^ Bengio, Yoshua; Mindermann, Sören; Privitera, Davide (29 January 2025). International AI Safety Report (Report). UK Department for Science, Innovation and Technology. Section 2.3: Systemic risks. DSIT 2025/001. Retrieved 17 November 2025.