Countering Adversarial Images Using Input Transformations: Difference between revisions

From statwiki
Jump to navigation Jump to search
No edit summary
Line 4: Line 4:




Hence an urgent need for approaches that increase the robustness of learning systems to such examples[[File:panda.jpg]]
Hence an urgent need for approaches that increase the robustness of learning systems to such examples[[File:Panda.jpg]]

Revision as of 18:15, 14 November 2018

Motivation

As the use of machine intelligence has increased , robustness has become a critical feature to guarantee the reliability of deployed machine-learning systems. However, recent research has shown that existing models are not robust to small , adversarial designed perturbations of the input. Adversarial examples are inputs to Machine Learning models that an attacker has intentionally designed to cause the model to make a mistake.The adversarial examples are not specific to Images , but also Malware, Text Understanding ,Speech. Below example, a small perturbation when applied to original image, the prediction is changed.


Hence an urgent need for approaches that increase the robustness of learning systems to such examplesFile:Panda.jpg