Timezone: »
FERMI: Fair Empirical Risk Minimization Via Exponential Rényi Mutual Information
Andrew Lowy · Rakesh Pavan · Sina Baharlouei · Meisam Razaviyayn · Ahmad Beirami
Despite the success of large-scale empirical risk minimization (ERM) at achieving high accuracy across a variety of machine learning tasks, fair ERM is hindered by the incompatibility of fairness constraints with stochastic optimization. In this paper, we propose the fair empirical risk minimization via exponential Rényi mutual information (FERMI) framework. FERMI is built on a stochastic estimator for exponential Rényi mutual information (ERMI), an information divergence measuring the degree of the dependence of predictions on sensitive attributes. Theoretically, we show that ERMI upper bounds existing popular fairness violation metrics, thus controlling ERMI provides guarantees on other commonly used violations, such as $L_\infty$. We derive an unbiased estimator for ERMI, which we use to derive the FERMI algorithm. We prove that FERMI converges for demographic parity, equalized odds, and equal opportunity notions of fairness in stochastic optimization. Empirically, we show that FERMI is amenable to large-scale problems with multiple (non-binary) sensitive attributes and non-binary targets. Extensive experiments show that FERMI achieves the most favorable tradeoffs between fairness violation and test accuracy across all tested setups compared with state-of-the-art baselines for demographic parity, equalized odds, equal opportunity. These benefits are especially significant for non-binary classification with large sensitive sets and small batch sizes, showcasing the effectiveness of the FERMI objective and the developed stochastic algorithm for solving it.
Author Information
Andrew Lowy (USC)
Rakesh Pavan (NITK)
Sina Baharlouei (UNIVERSITY OF SOUTHERN CALIFORNIA)
Meisam Razaviyayn (University of southern California)
Ahmad Beirami (Facebook AI)
More from the Same Authors
-
2022 : Improving adversarial robustness via joint classification and multiple explicit detection classes »
Sina Baharlouei · Fatemeh Sheikholeslami · Meisam Razaviyayn · Zico Kolter -
2023 Workshop: ICML2023 Neural Conversational AI Workshop - What’s left to TEACH (Trustworthy, Enhanced, Adaptable, Capable and Human-centric) chatbots? »
Hyundong Cho · Nayeon Lee · Ninareh Mehrabi · Hsuan Su · Ahmad Beirami · Hung-yi Lee · Jonathan May -
2022 Poster: A Rigorous Study of Integrated Gradients Method and Extensions to Internal Neuron Attributions »
Daniel Lundstrom · Tianjian Huang · Meisam Razaviyayn -
2022 Spotlight: A Rigorous Study of Integrated Gradients Method and Extensions to Internal Neuron Attributions »
Daniel Lundstrom · Tianjian Huang · Meisam Razaviyayn -
2021 Workshop: Information-Theoretic Methods for Rigorous, Responsible, and Reliable Machine Learning (ITR3) »
Ahmad Beirami · Flavio Calmon · Berivan Isik · Haewon Jeong · Matthew Nokleby · Cynthia Rush -
2021 : Opening Remarks »
Ahmad Beirami -
2021 Poster: Ditto: Fair and Robust Federated Learning Through Personalization »
Tian Li · Shengyuan Hu · Ahmad Beirami · Virginia Smith -
2021 Spotlight: Ditto: Fair and Robust Federated Learning Through Personalization »
Tian Li · Shengyuan Hu · Ahmad Beirami · Virginia Smith -
2018 Poster: Gradient Primal-Dual Algorithm Converges to Second-Order Stationary Solution for Nonconvex Distributed Optimization Over Networks »
Mingyi Hong · Meisam Razaviyayn · Jason Lee -
2018 Oral: Gradient Primal-Dual Algorithm Converges to Second-Order Stationary Solution for Nonconvex Distributed Optimization Over Networks »
Mingyi Hong · Meisam Razaviyayn · Jason Lee