Breakout Session [Hall 326A]
in
Affinity Workshop: 4th Women in Machine Learning (WiML) Un-Workshop
Deploying research and making real world impact
Fei Fang · Diyi Yang
Abstract:
There are many robustness issues arising in a variety of forms when deploying ML systems in the real world. For example, neural networks suffer from sensitivity to distributional shift, when a model is tested on a data distribution different from what it was trained on. Such a shift is frequently encountered in practical deployments and can lead to a substantial degradation in performance. In addition, neural networks are vulnerable to adversarial examples - small perturbations to the input can successfully fool classifiers into making incorrect predictions. In this section, we will develop a deeper understanding of different robustness issues and discuss how to effectively enhance models' robustness.
Chat is not available.