{"id":723253,"date":"2021-02-11T11:23:51","date_gmt":"2021-02-11T19:23:51","guid":{"rendered":"https:\/\/www.microsoft.com\/en-us\/research\/?p=723253"},"modified":"2021-03-24T13:17:46","modified_gmt":"2021-03-24T20:17:46","slug":"denoised-smoothing-provably-defending-pretrained-classifiers-against-adversarial-examples","status":"publish","type":"post","link":"https:\/\/www.microsoft.com\/en-us\/research\/blog\/denoised-smoothing-provably-defending-pretrained-classifiers-against-adversarial-examples\/","title":{"rendered":"Denoised smoothing: Provably defending pretrained classifiers against adversarial examples"},"content":{"rendered":"\n
\"An<\/figure>\n\n\n\n

Editor\u2019s note: This post and its research are the result of the collaborative efforts of a team of researchers comprising former Microsoft Research Engineer Hadi Salman (opens in new tab)<\/span><\/a>, CMU PhD student Mingjie Sun (opens in new tab)<\/span><\/a>, Researcher Greg Yang (opens in new tab)<\/span><\/a>, Partner Research Manager Ashish Kapoor (opens in new tab)<\/span><\/a>, and CMU Associate Professor J. Zico Kolter (opens in new tab)<\/span><\/a>.<\/em><\/p>\n\n\n\n

It\u2019s been well-documented that subtle modifications to the inputs of image classification systems can lead to bad predictions. Take, for example, a model trained to classify images of an elephant. The model easily classifies an image of the animal grazing in a grassy field. Now, if just a few pixels in that image are maliciously altered, you can get a very different\u2014and wrong<\/em>\u2014prediction despite the image appearing unchanged to the human eye. Sensitivity to such input perturbations, which are known as adversarial examples, raises security and reliability issues for the vision-based systems that we deploy in the real world. To tackle this challenge, recent research has revolved around building defenses against such adversarial examples. However, most of these adversarial defenses, such as randomized smoothing, require specifically training a classifier with a custom objective, which can be computationally expensive.<\/p>\n\n\n\n

\n\t