Countering Adversarial Images Using Input Transformations: Difference between revisions
Jump to navigation
Jump to search
No edit summary |
|||
Line 4: | Line 4: | ||
Hence an urgent need for approaches that increase the robustness of learning systems to such examples[[File:Panda. | Hence an urgent need for approaches that increase the robustness of learning systems to such examples[[File:Panda.png]] |
Revision as of 18:16, 14 November 2018
Motivation
As the use of machine intelligence has increased , robustness has become a critical feature to guarantee the reliability of deployed machine-learning systems. However, recent research has shown that existing models are not robust to small , adversarial designed perturbations of the input. Adversarial examples are inputs to Machine Learning models that an attacker has intentionally designed to cause the model to make a mistake.The adversarial examples are not specific to Images , but also Malware, Text Understanding ,Speech. Below example, a small perturbation when applied to original image, the prediction is changed.
Hence an urgent need for approaches that increase the robustness of learning systems to such examples