Skip to main content

Mr. Kaidi Xu

Northeastern University

Monday, March 15, 2021
3:00PM – 4:00PM – Zoom

Abstract

Deep learning has achieved extraordinary performance in many application domains recently. It has been well accepted
that Deep Neural Networks (DNNs) are vulnerable to adversarial attacks, which raises concerns of DNNs in securitycritical
applications and may result in disastrous consequences. Adversarial attacks are usually implemented by generating
adversarial examples, i.e., adding sophisticated perturbations onto benign examples, such that adversarial examples
are classified by the DNN as target (wrong) labels of the benign examples. Trustworthy machine learning aims to study
such phenomena and leverage them to build robust machine learning systems and explain DNNs.

In this talk, I will present the mechanism of trustworthy machine learning in both practical and theoretical ways. Specifically,
a uniform adversarial attack generation framework, structured attack (StrAttack) is introduced, which explores
group sparsity in adversarial perturbations by sliding a mask through images aiming for extracting key spatial structures.
Second, we discuss the feasibility of adversarial attacks in the physical world and introduce a convincing framework,
Expectation over Transformation (EoT). Utilize EoT with Thin Plate Spline (TPS) transformation, we can generate Adversarial
T-shirts, a powerful physical adversarial patch for evading person detectors in the real-world. Finally, we stand
on the defense side and introduce Linear relaxation-based perturbation analysis (LiRPA) for neural networks, which
computes provable linear bounds of output neurons given a certain amount of input perturbation. LiRPA studies the
adversarial example in a theoretical way and can guarantee the test accuracy of a model by given perturbation constraints.
The generality, flexibility, efficiency, and ease-of-use of the proposed framework facilitate the adoption of LiRPA
based provable methods for other machine learning problems beyond robustness verification.

For more info, please follow thisĀ link.