The evolution of AI capabilities has expanded the attack surface for malicious actors while simultaneously increasing the probability of unpredictable malfunctions within autonomous agents. The International AI Safety Report 2026 categorizes these emerging threats into three distinct yet overlapping domains: risks from malicious use, risks from malfunctions, and systemic risks. Malicious use is characterized by the intentional deployment of AI to cause harm, such as the generation of deepfakes for fraud, extortion, or political manipulation, and the acceleration of cyberattacks. Of particular concern is the qualitative leap in biological and chemical risks; for instance, the OpenAI o3 model has demonstrated the ability to outperform 94% of domain experts in troubleshooting complex virology lab protocols, effectively lowering the barrier for novices to develop dangerous biological agents.
Risks from malfunctions represent a second critical pillar, where AI systems operate outside intended parameters due to reliability challenges or a loss of control. Current systems frequently exhibit failures such as fabricating information, producing flawed code, and providing misleading medical advice. As AI agents gain greater autonomy, the ability for human operators to intervene before a failure causes significant harm is diminished. Finally, systemic risks involve broader societal harms, including labor market impacts where translation and writing jobs are increasingly substituted by AI, and the erosion of human autonomy through widespread reliance on automated decision-making.