Adversarial Attacks Against Reinforcement Learning Agents with Ian Goodfellow & Sandy Huang

Published: March 15, 2018, 4:27 p.m.

b'In this episode, I\\u2019m joined by Ian Goodfellow, Staff Research Scientist at Google Brain and Sandy Huang, Phd Student in the EECS department at UC Berkeley, to discuss their work on the paper Adversarial Attacks on Neural Network Policies. If you\\u2019re a regular listener here you\\u2019ve probably heard of adversarial attacks, and have seen examples of deep learning based object detectors that can be fooled into thinking that, for example, a giraffe is actually a school bus, by injecting some imperceptible noise into the image. Well, Sandy and Ian\\u2019s paper sits at the intersection of adversarial attacks and reinforcement learning, another area we\\u2019ve discussed quite a bit on the podcast. In their paper, they describe how adversarial attacks can also be effective at targeting neural network policies in reinforcement learning. Sandy gives us an overview of the paper, including how changing a single pixel value can throw off performance of a model trained to play Atari games. We also cover a lot of interesting topics relating to adversarial attacks and RL individually, and some related areas such as hierarchical reward functions and transfer learning. This was a great conversation that I\\u2019m really excited to bring to you! For complete show notes, head over to twimlai.com/talk/119'