Recommendations for algorithmic fairness assessments of predictive models in healthcare : evidence from large-scale empirical analyses

Placeholder Show Content

Abstract/Contents

Abstract
The use of machine learning to develop predictive models that inform clinical decision making has the potential to introduce and exacerbate health inequity. A growing body of work has framed these issues as ones of algorithmic fairness, seeking to develop techniques to anticipate and proactively mitigate harms. The central aim of my work is to provide and justify practical recommendations for the development and evaluation of clinical predictive models in alignment with these principles. Using evidence derived from large-scale empirical studies, I demonstrate that, when it is assumed that the predicted outcome is not subject to differential measurement error across groups and threshold selection is unconstrained, approaches that aim to incorporate fairness considerations into the learning objective used for model development typically do not improve model performance or confer greater net benefit for any of the studied patient populations compared to standard learning paradigms. For evaluation in this setting, I advocate for the use of criteria that assess the calibration properties of predictive models across groups at clinically-relevant decision thresholds. To contextualize the interplay between measures of model performance, fairness, and benefit, I present a case study for models that estimate the ten-year risk of atherosclerotic cardiovascular disease to inform statin initiation. Finally, I caution that standard observational analyses of algorithmic fairness in healthcare lack the contextual grounding and causal awareness necessary to reason about the mechanisms that lead to health disparities, as well as about the potential for technical approaches to counteract those mechanisms, and argue for refocusing algorithmic fairness efforts in healthcare on participatory design, transparent model reporting, auditing, and reasoning about the impact of model-enabled interventions in context.

Description

Type of resource text
Form electronic resource; remote; computer; online resource
Extent 1 online resource.
Place California
Place [Stanford, California]
Publisher [Stanford University]
Copyright date 2021; ©2021
Publication date 2021; 2021
Issuance monographic
Language English

Creators/Contributors

Author Pfohl, Stephen Robert
Degree supervisor Shah, Nigam
Thesis advisor Shah, Nigam
Thesis advisor Goel, Sharad, 1977-
Thesis advisor Hernandez-Boussard, Tina
Thesis advisor Zou, James
Degree committee member Goel, Sharad, 1977-
Degree committee member Hernandez-Boussard, Tina
Degree committee member Zou, James
Associated with Stanford University, Department of Biomedical Informatics

Subjects

Genre Theses
Genre Text

Bibliographic information

Statement of responsibility Stephen Pfohl.
Note Submitted to the Department of Biomedical Informatics.
Thesis Thesis Ph.D. Stanford University 2021.
Location https://purl.stanford.edu/xb296fk1005

Access conditions

Copyright
© 2021 by Stephen Robert Pfohl
License
This work is licensed under a Creative Commons Attribution 3.0 Unported license (CC BY).

Also listed in

Loading usage metrics...