How to fool neural networks with adversarial attacks

Speaker Grigory Malivenko  
Language EN
Stage Digital
Activity Presentation

Deep learning is a great approach to solve problems seemed unsolvable before and neural networks are now so widespread that they are used almost everywhere - from mobile applications to safety-critical environments. However, deep neural networks have been (almost) recently found vulnerable to well-designed input samples, called adversarial samples. Adversarial samples may be imperceptible to human but can easily fool deep neural networks in the testing/deploying stage.

In this presentation I'll show how to perform adversarial attack to pretrained model with known architecture and how to make your network less vulnerable to this kind of attack.

Requirements for attendees

basic computer science knowledge