Countering Adversarial Images Using Input Transformations: Difference between revisions
No edit summary |
No edit summary |
||
Line 1: | Line 1: | ||
== Motivation == | == Motivation == | ||
As the use of machine intelligence has increased , robustness has become a critical feature to guarantee the reliability of deployed machine-learning systems. However, recent research has shown that existing models are not robust to small , adversarial designed perturbations of the input. Adversarial examples are inputs to Machine Learning models that an attacker has intentionally designed to cause the model to make a mistake.The adversarial examples are not specific to Images , but also Malware, Text Understanding ,Speech. | As the use of machine intelligence has increased , robustness has become a critical feature to guarantee the reliability of deployed machine-learning systems. However, recent research has shown that existing models are not robust to small , adversarial designed perturbations of the input. Adversarial examples are inputs to Machine Learning models that an attacker has intentionally designed to cause the model to make a mistake.The adversarial examples are not specific to Images , but also Malware, Text Understanding ,Speech. | ||
Below example, a small perturbation when applied to original image, the prediction is changed. | Below example (Goodfellow et. al), a small perturbation when applied to original image of panda, the prediction is changed to gibbon. | ||
[[File:Panda.png]] | |||
Hence an urgent need for approaches that increase the robustness of learning systems to such examples | |||
Hence an urgent need for approaches that increase the robustness of learning systems to such examples |
Revision as of 18:18, 14 November 2018
Motivation
As the use of machine intelligence has increased , robustness has become a critical feature to guarantee the reliability of deployed machine-learning systems. However, recent research has shown that existing models are not robust to small , adversarial designed perturbations of the input. Adversarial examples are inputs to Machine Learning models that an attacker has intentionally designed to cause the model to make a mistake.The adversarial examples are not specific to Images , but also Malware, Text Understanding ,Speech. Below example (Goodfellow et. al), a small perturbation when applied to original image of panda, the prediction is changed to gibbon.
Hence an urgent need for approaches that increase the robustness of learning systems to such examples