Timezone: »
Machine learning models are often personalized with categorical attributes that define groups. In this work, we show that personalization with group attributes can inadvertently reduce performance at a group level -- i.e., groups may receive unnecessarily inaccurate predictions by sharing their personal characteristics. We present formal conditions to ensure the fair use of group attributes in a prediction task, and describe how they can be checked by training one additional model. We characterize how fair use conditions be violated due to standard practices in model development, and study the prevalence of fair use violations in clinical prediction tasks. Our results show that personalization often fails to produce a tailored performance gain for every group who reports personal data, and underscore the need to evaluate fair use when personalizing models with characteristics that are protected, sensitive, self-reported, or costly to acquire.
Author Information
Vinith Suriyakumar (Massachusetts Institute of Technology)
Marzyeh Ghassemi (MIT)

Dr. Marzyeh Ghassemi is an Assistant Professor at MIT in Electrical Engineering and Computer Science (EECS) and Institute for Medical Engineering & Science (IMES), and a Vector Institute faculty member holding a Canadian CIFAR AI Chair and Canada Research Chair. She holds MIT affiliations with the Jameel Clinic and CSAIL. Professor Ghassemi holds a Herman L. F. von Helmholtz Career Development Professorship, and was named a CIFAR Azrieli Global Scholar and one of MIT Tech Review’s 35 Innovators Under 35. Previously, she was a Visiting Researcher with Alphabet’s Verily. She is currently on leave from the University of Toronto Departments of Computer Science and Medicine. Prior to her PhD in Computer Science at MIT, she received an MSc. degree in biomedical engineering from Oxford University as a Marshall Scholar, and B.S. degrees in computer science and electrical engineering as a Goldwater Scholar at New Mexico State University.
Berk Ustun (UCSD)
Related Events (a corresponding poster, oral, or spotlight)
-
2023 Oral: When Personalization Harms Performance: Reconsidering the Use of Group Attributes in Prediction »
Thu. Jul 27th 02:00 -- 02:08 AM Room Ballroom C
More from the Same Authors
-
2021 : Learning through Recourse under Censoring »
Jennifer Chien · Berk Ustun -
2022 : Evaluating and Improving Robustness of Self-Supervised Representations to Spurious Correlations »
Kimia Hamidieh · Haoran Zhang · Marzyeh Ghassemi -
2023 : Prediction without Preclusion: Recourse Verification with Reachable Sets »
Avni Kothari · Bogdan Kulynych · Lily Weng · Berk Ustun -
2023 : Prediction without Preclusion Recourse Verification with Reachable Sets »
Avni Kothari · Berk Ustun · Lily Weng · Bogdan Kulynych -
2023 : Identifying Implicit Social Biases in Vision-Language Models »
Kimia Hamidieh · Haoran Zhang · Thomas Hartvigsen · Marzyeh Ghassemi -
2023 : Participatory Personalization in Classification »
Hailey Joren · Chirag Nagpal · Katherine Heller · Berk Ustun -
2023 : Prediction without Preclusion Recourse Verification with Reachable Sets »
Avni Kothari · Bogdan Kulynych · Lily Weng · Berk Ustun -
2023 : Participatory Personalization in Classification »
Hailey Joren · Chirag Nagpal · Katherine Heller · Berk Ustun -
2023 : A Pipeline for Interpretable Clinical Subtyping with Deep Metric Learning »
Haoran Zhang · Qixuan Jin · Thomas Hartvigsen · Miriam Udler · Marzyeh Ghassemi -
2023 : A Pipeline for Interpretable Clinical Subtyping with Deep Metric Learning »
Haoran Zhang · Qixuan Jin · Thomas Hartvigsen · Miriam Udler · Marzyeh Ghassemi -
2023 : Participatory Personalization in Classification »
Hailey Joren · Chirag Nagpal · Katherine Heller · Berk Ustun -
2023 : Identifying Implicit Social Biases in Vision-Language Models »
Kimia Hamidieh · Haoran Zhang · Thomas Hartvigsen · Marzyeh Ghassemi -
2023 : The pulse of ethical machine learning and health »
Marzyeh Ghassemi -
2023 Poster: Change is Hard: A Closer Look at Subpopulation Shift »
Yuzhe Yang · Haoran Zhang · Dina Katabi · Marzyeh Ghassemi -
2023 Poster: "Why did the Model Fail?": Attributing Model Performance Changes to Distribution Shifts »
Haoran Zhang · Harvineet Singh · Marzyeh Ghassemi · Shalmali Joshi -
2023 Invited Talk: Taking the Pulse Of Ethical ML in Health »
Marzyeh Ghassemi -
2022 : Invited talks 2 Q/A, Christina and Marzyeh »
Christina Heinze-Deml · Marzyeh Ghassemi -
2022 : Invited talks 2, Christina Heinze-Deml and Marzyeh Ghassemi »
Christina Heinze-Deml · Marzyeh Ghassemi -
2022 Poster: Public Data-Assisted Mirror Descent for Private Model Training »
Ehsan Amid · Arun Ganesh · Rajiv Mathews · Swaroop Ramaswamy · Shuang Song · Thomas Steinke · Thomas Steinke · Vinith Suriyakumar · Om Thakkar · Abhradeep Guha Thakurta -
2022 Spotlight: Public Data-Assisted Mirror Descent for Private Model Training »
Ehsan Amid · Arun Ganesh · Rajiv Mathews · Swaroop Ramaswamy · Shuang Song · Thomas Steinke · Thomas Steinke · Vinith Suriyakumar · Om Thakkar · Abhradeep Guha Thakurta -
2021 Poster: Simultaneous Similarity-based Self-Distillation for Deep Metric Learning »
Karsten Roth · Timo Milbich · Bjorn Ommer · Joseph Paul Cohen · Marzyeh Ghassemi -
2021 Spotlight: Simultaneous Similarity-based Self-Distillation for Deep Metric Learning »
Karsten Roth · Timo Milbich · Bjorn Ommer · Joseph Paul Cohen · Marzyeh Ghassemi -
2020 : Panel 2 »
Deborah Raji · Berk Ustun · Alexandra Chouldechova · Jamelle Watson-Daniels -
2020 : Actionable Recourse in Machine Learning »
Berk Ustun -
2020 Workshop: Healthcare Systems, Population Health, and the Role of Health-tech »
Creighton Heaukulani · Konstantina Palla · Katherine Heller · Niranjani Prasad · Marzyeh Ghassemi -
2020 Poster: Predictive Multiplicity in Classification »
Charles Marx · Flavio Calmon · Berk Ustun -
2019 Poster: Fairness without Harm: Decoupled Classifiers with Preference Guarantees »
Berk Ustun · Yang Liu · David Parkes -
2019 Poster: Repairing without Retraining: Avoiding Disparate Impact with Counterfactual Distributions »
Hao Wang · Berk Ustun · Flavio Calmon -
2019 Oral: Repairing without Retraining: Avoiding Disparate Impact with Counterfactual Distributions »
Hao Wang · Berk Ustun · Flavio Calmon -
2019 Oral: Fairness without Harm: Decoupled Classifiers with Preference Guarantees »
Berk Ustun · Yang Liu · David Parkes