Campuses:

security

Monday, April 22, 2019 - 1:25pm - 2:25pm
Tom Goldstein (University of Maryland)
Neural networks solve complex computer vision problems with human-like accuracy. However, it has recently been observed that neural nets are easily fooled and manipulated by adversarial examples, in which an attacker manipulates the network by making tiny changes to its inputs. In this talk, I give a high-level overview of adversarial examples, and then discuss a newer type of attack called data poisoning, in which a network is manipulated at train time rather than test time.
Subscribe to RSS - security