University of Cambridge > Talks.cam > Engineering Safe AI > An introduction to adversarial attacks and defences

An introduction to adversarial attacks and defences

Download to your calendar using vCal

If you have a question about this talk, please contact Adrià Garriga Alonso .

AI safety is not limited to RL settings. For example, we can use machine learning algorithms to design spam filters, yet attackers can still “reverse-engineer” our defence to send us junk emails. Autonomous driving systems based on computer vision techniques are also vulnerable to attacks, for instance, attackers can carefully apply a sticker to a stop sign in order to fool the vision system of the car. In this talk I will briefly discuss the mathematical framework of these attack techniques (specifically on image classifiers) and defence techniques against them.

Slides available here: http://yingzhenli.net/home/pdf/attack_defence.pdf

This talk is part of the Engineering Safe AI series.

This talk is included in these lists:

Note that ex-directory lists are not shown.

 

© 2006-2025 Talks.cam, University of Cambridge. Contact Us | Help and Documentation | Privacy and Publicity