{"id":138539,"date":"2025-04-24T09:00:00","date_gmt":"2025-04-24T16:00:00","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/security\/blog\/?p=138539"},"modified":"2025-04-29T09:09:10","modified_gmt":"2025-04-29T16:09:10","slug":"new-whitepaper-outlines-the-taxonomy-of-failure-modes-in-ai-agents","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/security\/blog\/2025\/04\/24\/new-whitepaper-outlines-the-taxonomy-of-failure-modes-in-ai-agents\/","title":{"rendered":"New whitepaper outlines the taxonomy of failure modes in AI agents"},"content":{"rendered":"\n
We are releasing a taxonomy of failure modes in AI agents<\/a> to help security professionals and machine learning engineers think through how AI systems can fail and design them with safety and security in mind.<\/p>\n\n\n\n The taxonomy continues Microsoft AI Red Team’s work to lead the creation of systematization of failure modes in AI; in 2019, we published<\/a> one of the earliest industry efforts enumerating the failure modes of traditional AI systems. In 2020, we partnered with MITRE and 11 other organizations to codify the security failures in AI systems<\/a> as Adversarial ML Threat Matrix, which has now evolved into MITRE ATLAS\u2122. This effort is another step in helping the industry think through what the safety and security failures in the fast-moving and highly impactful agentic AI space are.<\/p>\n\n\n