Skip to main content
Patterns logoLink to Patterns
. 2024 Sep 12;5(10):101059. doi: 10.1016/j.patter.2024.101059

FAIM: Fairness-aware interpretable modeling for trustworthy machine learning in healthcare

Mingxuan Liu 1, Yilin Ning 1, Yuhe Ke 2, Yuqing Shang 1, Bibhas Chakraborty 1,3,4,5, Marcus Eng Hock Ong 3,6, Roger Vaughan 1,3, Nan Liu 1,3,7,8,
PMCID: PMC11573921  PMID: 39569213

Summary

The escalating integration of machine learning in high-stakes fields such as healthcare raises substantial concerns about model fairness. We propose an interpretable framework, fairness-aware interpretable modeling (FAIM), to improve model fairness without compromising performance, featuring an interactive interface to identify a “fairer” model from a set of high-performing models and promoting the integration of data-driven evidence and clinical expertise to enhance contextualized fairness. We demonstrate FAIM’s value in reducing intersectional biases arising from race and sex by predicting hospital admission with two real-world databases, the Medical Information Mart for Intensive Care IV Emergency Department (MIMIC-IV-ED) and the database collected from Singapore General Hospital Emergency Department (SGH-ED). For both datasets, FAIM models not only exhibit satisfactory discriminatory performance but also significantly mitigate biases as measured by well-established fairness metrics, outperforming commonly used bias mitigation methods. Our approach demonstrates the feasibility of improving fairness without sacrificing performance and provides a modeling mode that invites domain experts to engage, fostering a multidisciplinary effort toward tailored AI fairness.

Keywords: AI fairness, fairness in healthcare, interpretable machine learning, trustworthy machine learning

Graphical abstract

graphic file with name fx1.jpg

Highlights

  • FAIM framework to tackle bias in healthcare

  • FAIM improves model fairness while maintaining performance via near optimality

  • FAIM features an interactive interface to enhance contextualized fairness

  • FAIM links interpretability with fairness to verify the effects of bias mitigation

The bigger picture

Fairness in clinical decision-making is a vital aspect of healthcare systems, and in the development of artificial intelligence (AI) algorithms, a lack of fairness can perpetuate existing biases and potentially harm patients. AI fairness aims to address and mitigate health disparities exacerbated by algorithmic decisions, promoting equitable healthcare delivery. However, the complex nature of clinical settings poses unique challenges, such as uncertainty in using non-clinical variables, addressing biases in clinical variables, and the severe consequences of wrong decisions. This complexity underscores the need for interpretability and domain expert involvement to contextualize AI fairness to specific clinical scenarios. The fairness-aware interpretable modeling (FAIM) framework assists in the selection of fairness-aware models in clinical settings while maintaining model performance.


This study introduces FAIM, a fairness-aware and interpretable modeling framework designed to build well-performing yet fair models as alternatives to performance-optimized, fairness-unaware models. FAIM offers a new approach to bias mitigation by anchoring the performance of fairness-unaware models, exploring fairness diversity among nearly optimal models, and achieving contextualized fairness. The study shows that fairness can be improved without compromising performance, and by enhancing interpretability and fostering active engagement with domain experts, FAIM promotes contextualized AI fairness, encouraging clinician involvement and interdisciplinary collaboration.

Introduction

As artificial intelligence (AI) gains prominence in high-stakes fields like healthcare, concerns about fairness have grown.1,2,3 Biases (as opposed to fairness) in machine learning arise when “sensitive” factors (e.g., age, sex, gender, race, ethnicity, socio-economic status, etc.) unjustly skew decision-making.4,5 In healthcare, a biased model can unjustly influence life-altering decisions, such as disease diagnosis6,7 and organ allocation.8 Clinical AI fairness is challenging due to healthcare’s complexity and the impact of social determinants, and its integration into clinical decision-making is crucial to prevent the escalation of health disparities.9,10,11

A seemingly straightforward approach to reduce bias is to simply exclude sensitive variables from the decision-making process, also named “under blindness,” but this has been deemed undesirable, particularly in cases where subpopulations (such as by sex) are distinct.12,13 Numerous methods have been developed to systematically mitigate biases, categorized by the stage of the modeling process at which they operate.14,15 Bias mitigation can occur pre-process through data adjustment (e.g., sampling16 and reweighing16), in-process via direct fair model development (e.g., regularization17,18,19 and representation learning20,21), or post-process by altering model outputs (e.g., equalized-odds post-process22). However, pre-process methods often encounter challenges when addressing biases that involve intersecting multiple attributes (e.g., race and sex).23,24 Post-process approaches, by modifying outputs, fail to address the root causes of biases, leaving predictions altered but not clarified.25

Fairness in machine learning is hindered by a lack of interpretability, especially those employing models with black-box architectures for bias mitigation. Interpretability concerns also arise when randomness is utilized to post-process individual predictions (e.g., changing a positive prediction into a negative one) for the purpose of group fairness without clear clinical justification.22,26,27 In healthcare, the contextual nature of fairness necessitates the specialized knowledge of clinicians to ensure that fairness definitions align with clinical realities, as AI fairness and its decision-making consequences differ markedly across medical fields.11,12 However, the lack of interpretability poses formidable obstacles to their active participation in the modeling process, challenging the mutual understanding between AI developers and clinicians and impeding the translation of technical advancements into clinical practice.

Another common limitation of existing bias mitigation methods is that they often compromise model performance, as measured by machine learning metrics like area under the receiver operating characteristic (ROC) curve (AUC). This drawback can hamper practical application, cast doubt on the claimed fairness due to increased uncertainty, and potentially lead to severe and unforeseen consequences. According to a recent empirical study,28 current bias mitigation methods degraded model performance in approximately 50% cases, whereas in 25% cases, they worsened both fairness and performance. Specifically, Pfohl et al.29 empirically demonstrated that the methods penalizing prediction discrepancies (which are often interpreted as evidence of bias) can almost universally decrease multiple aspects of performance. Nevertheless, it has been empirically shown that such a trade-off between model performance and fairness is not inevitable.5,27

To bridge the gap in clinical AI fairness, we propose a new fairness-aware interpretable modeling (FAIM) framework (Figure 1) to achieve fairness with enhanced interpretability without sacrificing model performance. FAIM operates within a set of nearly optimal models that offer high performance without necessarily reaching optimality.30 By leveraging the varying degrees of model reliance on variables (including sensitive variables) within the set,31,32 FAIM can identify alternative model formulations that improve model fairness without significantly impairing performance. Rather than producing a single final model and leaving uncertainty about which aspects of fairness are compromised for performance, FAIM utilizes models’ diverse fairness profiles and near-optimal performance to facilitate informed discussions between clinicians and model developers to prioritize fair models for specific contexts. FAIM also examines the impact of excluding some (or all) sensitive variables on model fairness, visualizes the findings to help clinicians select a fairness-enhanced model with reasonable interpretation and contextualization, and employs explainable AI (XAI) methods to clarify variable importance changes due to the fairness enhancement to further improve interpretation. We illustrate our method in the prediction of hospital admission in the emergency department (ED) using two large-scale clinical datasets, focusing on reducing potential bias due to two sensitive variables, i.e., race and sex. Although we use clinical case studies, the framework introduced has broader applicability across multiple domains and diverse tasks.

Figure 1.

Figure 1

The general framework of FAIM

The FAIM framework consists of three modules to locate a fairer model in a set of nearly optimal models: nearly optimal model generation, model selection with contextualized fairness, and model explanation.

Results

We implemented FAIM on two large clinical datasets to predict hospital admission: the Medical Information Mart for Intensive Care IV ED (MIMIC-IV-ED)33 and data collected from the Singapore General Hospital ED (SGH-ED),34 with the characteristics summarized in Tables S1 and S2. We aimed to mitigate biases related to sex and race, denoted by Xs, and other variables were considered less sensitive, denoted as XU, due to their uncertain roles regarding race and sex biases. Particularly, the variable of age was considered less sensitive because of its known high predictive ability for hospital admission. We used logistic regression as the fairness-unaware baseline model ("baseline” for short), aligning with clinical studies that value not only the discrimination performance but also model transparency and statistical functionality (e.g., odds ratios for predictors).35 As demonstrated in the two datasets, the FAIM framework transparently navigated the exclusion of sensitive variables, and the FAIM output models (i.e., fairness-aware models, “FAIM” for short) had satisfactory discriminatory performance and significantly improved model fairness compared to other commonly used bias mitigation methods.

Model selection among nearly optimal models

After generating a set of nearly optimal models (see details in subsection nearly optimal model generation in the experimental procedures), we located the fairness-aware model via the interactive interface of FAIM. For the SGH-ED data, Figure 2 shows the distribution of coefficients for the nearly optimal models, showcasing various spectra of coefficients for sensitive variables sex and race as well as less-sensitive variables such as temperature and triage score (i.e., patient acuity category scale [PACS] used in Singapore). Such variation in coefficients suggests that while model performance is not impacted (Figure 2A), these nearly optimal models respond differently to variables, providing room for fairness improvement. Figure 3 showcases the fairness distribution of these nearly optimal models, aiding in model selection with data-driven evidence. In Figure 3A, regarding individual fairness metrics (equalized odds, equal opportunity, and balanced error rate [BER], as defined in Table 1), these nearly optimal models displayed diverse fairness profiles. Across panels, these models were jointly assessed and ranked using the proposed fairness ranking index (FRI), which aggregates these individual fairness metrics for a comprehensive assessment of fairness (see the experimental procedures for a detailed definition). As visualized in Figure 3B and quantified in Figure 3C, among the top 10 models, the top 5 models excluded race, whereas the sixth-ranked model excluded both race and sex (shown in the golden box); conversely, none of the models in the bottom 100 excluded race (shown in the gray box), which suggested the necessity of excluding race. We selected the top 1 model (i.e., model ID 224) as the FAIM output, which included the sex variable. This decision was informed by the similar distribution of nearly optimal models regardless of sex variable exclusion, indicating minimal bias caused by sex (Figure 3C), coupled with the absence of clinically detected sex biases in Singapore’s hospital admissions so far.

Figure 2.

Figure 2

Distribution of model performance and coefficient variability for nearly optimal models (n = 800) on SGH-ED data

(A) Histogram showing the distribution of model performance (AUC) for nearly optimal models evaluated on the validation set.

(B) Violin plot depicting the variability of model coefficients across different variables. The width of each violin represents the range of coefficient values, while the color gradient indicates the level of model performance, with darker shades corresponding to higher performance. The variables PACS triage score, temperature, race, and sex display wide spectra of coefficients.

Figure 3.

Figure 3

The interactive plot for model selection based on the nearly optimal models’ fairness evaluated on the validation set of SGH-ED data

(A) The graphical interface ranks nearly optimal models based on fairness metrics—equalized odds, equal opportunity, and BER equality. Model 224 is ranked first by the fairness ranking index (FRI), highlighted as the default fairness-aware model. Users can interactively engage with the data by hovering over points to display model details, zooming in on sections like the top 10 models, and adjusting panel views.

(B) Illustration of the panel of equalized odds metric with top 10 models showcased in the golden box, showing four models (stars) excluding both sex and race and six models (circles) excluding only race. The bottom 100 models, showcased in the gray box, predominantly include no exclusions (rectangles), with a minority excluding sex (upward triangles).

(C) A tabulation of models based on the exclusion cases of sensitive variable(s) detailing the counts of models in different fairness ranges—from “most fair” to “least fair.” The rightmost column records the highest ranking obtained by the nearly optimal models for each exclusion case.

Table 1.

Separation-based fairness metrics: Formulas and descriptions

Fairness metricsa Formulas Description
Equalized odds max{RangeS(P[Yˆ=1|Y=1,XS]),RangeS(P[Yˆ=1|Y=0,XS])} this metric checks if both true positive rates and false positive rates are consistent among subgroups, indicating that the likelihood of a correct or incorrect prediction does not vary significantly among subgroups
Equalized opportunity RangeS(P[Yˆ=1|Y=1,XS]) this metric assesses if the true positive rates (sensitivity) are consistent among subgroups, ensuring that the model’s ability to correctly identify positive cases is consistent among subgroups
Balanced error rate equality RangeS(P[Yˆ=1|Y=0,XS]+P[Yˆ=0|Y=1,XS]) this metric evaluates if the balanced error rate, which is the sum of the false positive rate and the false negative rate, is similar across subgroups, indicating balanced misclassification rates among subgroups
a

These three metrics are limited to binary outcomes. The smaller values indicate better fairness and fewer biases.

For the MIMIC-IV-ED dataset, Figure S1 shows the various spectra of coefficients for sensitive variables and less-sensitive variables, such as temperature and triage score (i.e., emergency severity index [ESI] used in the US). FAIM’s default choice was the top 1 model (i.e., model ID 681) with both race and sex excluded. This preference aligned with nine out of the top 10 models, as depicted in Figure S2. Similarly, we retained the top 1 model as the FAIM output model, reflecting the majority preference of the top-ranking models and addressing known racial and sex disparities in ED care outcomes in the US.36,37,38

Evaluation of model fairness

With the FAIM models above, we evaluated model fairness on the split-out test set. FAIM consistently outperformed both baseline models (i.e., fairness-unaware models) and other bias mitigation methods in model fairness except for the reductions method (see Table 2 for MIMIC-IV-ED data and Table 3 for SGH-ED data). FAIM’s improvement in fairness was statistically significant (p < 0.001), yielding a 53.5%–57.6% enhancement in fairness metrics for the MIMIC-IV-ED case and a 17.7%–21.7% enhancement for the SGH-ED case, compared with baseline models. Importantly, while the reductions method did improve fairness, it came at the cost of prediction performance, which led to impartially poor performance across different subgroups (more details in the next subsection, evaluation of model performance and statistical functionality).

Table 2.

Two-dimension evaluation of model fairness and performance for bias mitigation methods using MIMIC-IV-ED data

Fairness metricsa
Performance metrics
Equal opportunity Equalized odds BER equalityb AUC Sensitivityc Specificityc
Baselined 0.316 0.316 0.301 0.790 [0.787, 0.793] 0.715 [0.711, 0.720] 0.724 [0.720, 0.728]
Under blindnesse 0.159 0.159 0.155 0.787 [0.784, 0.790] 0.707 [0.703, 0.711] 0.729 [0.724, 0.733]
FAIM 0.134 0.147 0.140 0.786 [0.783, 0.789] 0.725 [0.720, 0.729] 0.711 [0.707, 0.715]
Reweighing (pre-process) 0.206 0.206 0.183 0.789 [0.786, 0.792] 0.709 [0.704, 0.713] 0.731 [0.726, 0.735]
Reductions (in-process) 0.037 0.037 0.032 0.706 [0.703, 0.709] 0.657 [0.652, 0.662] 0.755 [0.751, 0.759]
Fair-GLM (in-process) 0.146 0.151 0.148 0.786 [0.783, 0.789] 0.719 [0.714, 0.723] 0.715 [0.711, 0.720]
Equalized-odds post-processing (post-process) 0.734 0.734 0.554 0.594 [0.591, 0.596] 0.335 [0.330, 0.340] 0.852 [0.849, 0.855]
Adnetf (in process) 0.158 0.189 0.174 0.743 [0.740, 0.747] 0.703 [0.699, 0.708] 0.660 [0.656, 0.664]
a

Smaller values indicate higher levels of fairness.

b

BER: equality of the combination of the true positive rate and the true negative rate.

c

The thresholds were determined by Youden’s J index for methods that can yield predictive probabilities (i.e., original logistics regression, FAIM, reweighing, Fair-GLM, Adnet, and under blindness). The in-process method reductions and the post-process method equalized-odds post-process directly generated the binary prediction.

d

Baseline: the original logistics regression model, i.e., fairness-unaware baseline model.

e

Under blindness: the logistics regression with sensitive variables excluded.

f

Adnet: bias mitigation method utilizing adversarial learning, i.e., black-box models, instead of manipulating regression models to achieve fairness.

Table 3.

Two-dimension evaluation of model fairness and performance for bias mitigation methods using SGH-ED data

Fairness metricsa
Performance metrics
Equal opportunity Equalized odds BER equalityb AUC Sensitivityc Specificityc
Baselined 0.299 0.299 0.237 0.804 [0.803, 0.806] 0.713 [0.710, 0.715] 0.753 [0.751, 0.755]
Under blindnesse 0.252 0.252 0.209 0.803 [0.802, 0.805] 0.720 [0.717, 0.722] 0.744 [0.742, 0.746]
FAIM 0.234 0.234 0.195 0.802 [0.801, 0.804] 0.719 [0.717, 0.721] 0.746 [0.744, 0.748]
Reweighing (pre-process) 0.244 0.244 0.212 0.803 [0.802, 0.805] 0.712 [0.710, 0.715] 0.750 [0.748, 0.752]
Reductions (in-process) 0.035 0.036 0.036 0.664 [0.662, 0.665] 0.542 [0.540, 0.545] 0.785 [0.784, 0.787]
Fair-GLM (in-process) 0.260 0.260 0.216 0.802 [0.801, 0.804] 0.716 [0.713, 0.718] 0.747 [0.745, 0.748]
Equalized-odds post-processing (post-process) 0.685 0.685 0.490 0.601 [0.600, 0.603] 0.327 [0.323, 0.330] 0.875 [0.874, 0.876]
Adnetf (in-process) 0.357 0.357 0.337 0.748 [0.747, 0.750] 0.695 [0.693, 0.698] 0.675 [0.673, 0.677]
a

Smaller values indicate higher levels of fairness.

b

BER equality: equality of the combination of the true positive rate and the true negative rate.

c

The thresholds were determined by Youden’s J index for methods that can yield predictive probabilities (i.e., original logistics regression, FAIM, reweighing, Fair-GLM, Adnet, and under blindness). The in-process method reductions and the post-process method equalized-odds post-process directly generated the binary prediction.

d

Baseline: the original logistics regression model, i.e., fairness-unaware baseline model.

e

Under blindness: the logistics regression with sensitive variables excluded.

f

Adnet: bias mitigation method utilizing adversarial learning, i.e., black-box models, instead of manipulating regression models to achieve fairness.

Our ablation study suggests that FAIM’s advantages extend well beyond merely omitting sensitive variables, referred to as the under-blindness method. Indeed, regarding fairness metrics, FAIM’s performance improvement over under blindness was notable—ranging from 7.55% to 15.72% for the MIMIC-IV-ED dataset (Table 2) and from 6.69% to 7.14% for the SGH-ED dataset (Table 3).

Beyond fairness metrics, we conducted subgroup analyses by sensitive variables to evaluate FAIM’s impact on fairness. Table 4 shows that FAIM substantially reduced disparities in both true positive and true negative rates across racial subgroups for both datasets. For sex subgroups, FAIM notably reduced the disparities for the MIMIC-IV-ED dataset, while its effect on the SGH-ED dataset was not consistently significant, given the relatively small initial disparities present in the baseline model.

Table 4.

Subgroup analysis regarding sex and race

MIMIC-IV-ED
SGH-ED
Baseline Under blindnessa FAIM Baseline Under blindness FAIM
Race ΔTPRb 0.251 0.133 0.122c 0.264 0.223 0.211c
Race ΔTNRd 0.224 0.130 0.129c 0.170 0.149 0.134c
Sex ΔTPR 0.074 0.025 0.018c 0.015 0.001c 0.005
Sex ΔTNR 0.086 0.034 0.027c 0.008 0.025c 0.027
a

Under blindness: the logistics regression with sensitive variables excluded.

b

ΔTPR: the gap of true positive rate among race/ethnicity or sex subgroups.

c

Method that yields the least disparity.

d

ΔTNR: the gap of true negative rate among race/ethnicity or sex subgroups.

To further explore the changes in model reliance on the variables that lead to the improvement in fairness by FAIM, we analyzed this by comparing the variable importance between the fairness-unaware (i.e., baseline) and fairness-aware (i.e., FAIM) models, measured by Shapley additive explanations (SHAP).39 Beyond the sensitive variables of sex and race, for both datasets, the FAIM models preserved the variable contributions of less-sensitive variables, similar to the baseline models, while making minor changes to address fairness. As indicated by SHAP (Figure 4), the FAIM models preserved the importance of triage scores—ESI for MIMIC-IV-ED and PACS for SGH-ED—but downplayed the “pain scale” in the MIMIC-IV-ED case (Figure 4A).

Figure 4.

Figure 4

Variable importance analysis in both fairness-unaware (i.e., baseline) and fairness-aware (i.e., FAIM) models

(A) Results based on MIMIC-IV-ED dataset. (B) Results based on SGH-ED dataset. The fairness-aware model refers to the ranked no. 1 fairest model yielded by FAIM, and the fairness-unaware model refers to the baseline model, i.e., the logistics regression model.

Evaluation of model performance and statistical functionality

With improved model fairness, we then evaluated FAIM from the perspective of model performance. Our analysis of classification metrics—AUC, sensitivity, and specificity, detailed in Tables 2 and 3—revealed that FAIM maintained a comparable performance to the baseline models. For the MIMIC-IV-ED dataset, FAIM achieved an AUC of 0.786 (95% confidence interval [CI]: 0.783–0.789), which closely aligned with the baseline model’s AUC of 0.790 (95% CI: 0.787–0.793). Similarly, in the SGH-ED dataset, FAIM’s AUC was 0.802 (95% CI: 0.801–0.804), which closely matched the baseline model’s AUC of 0.804 (95% CI: 0.803–0.806).

The sensitivity and specificity values achieved by FAIM were also on par with the baseline models. In the MIMIC-IV-ED data, FAIM achieved a sensitivity of 0.725 (95% CI: 0.720–0.729) and a specificity of 0.711 (95% CI: 0.707–0.715), compared to the baseline model with a sensitivity of 0.715 (95% CI: 0.711–0.720) and a specificity of 0.724 (95% CI: 0.720–0.728). In the SGH-ED data, FAIM delivered a sensitivity of 0.719 (95% CI: 0.717–0.721) and a specificity of 0.746 (95% CI: 0.744–0.748), while the baseline model had a sensitivity of 0.713 (95% CI: 0.710–0.715) and a specificity of 0.753 (95% CI: 0.751–0.755). Among other bias mitigation methods, in-process methods such as reductions and post-process methods such as equalized-odds post-process heavily degraded the classification performance.

FAIM models’ odds ratios, as well as the corresponding statistical significance, were closely aligned with the fairness-unaware baseline models, with minor changes addressing fairness (Figure 5). This contrasts with some conventional bias mitigation methods, especially those based on black-box models, which often compromise statistical functionality. The sensitive variables race and/or sex, which were statistically significant in the baseline models, were automatically excluded owing to their minimal effects on model predictive ability. Specifically, in the case of MIMIC-IV-ED data, the pain scale variable became less significant in FAIM (Figure 5), aligning with its reduced variable importance as shown in SHAP analyses (Figure 4).

Figure 5.

Figure 5

Comparison of the odds ratios between fairness-unaware (i.e., baseline) and fairness-aware (i.e., FAIM) models

(A) Results based on SGH-ED dataset. (B) Results based on MIMIC-IV-ED dataset. The framework of FAIM excluded both sensitive variables race and sex from modeling to address fairness. As a result, the FAIM model did not have odds ratios for these two variables. Similarly, on the SGH-ED dataset, the variable race was excluded.

In terms of runtime efficiency, FAIM requires moderate training time—more than simpler methods like reweighing and under blindness, comparable to regularization-based methods and less than representation-learning-based methods due to the absence of hyperparameter tuning (see Table S3).

Discussion

The pursuit of AI fairness is of increasing importance, particularly in high-stakes decision-making.1,2,3 We contribute to this field by introducing FAIM, a fairness-aware, interpretable framework, to build well-performing yet fair models that can serve as alternatives to the fairness-unaware model that is solely optimized for performance. We provide a different mindset for bias mitigation, that is, to anchor the performance achieved by the fairness-unaware model, explore the diversity of fairness among the nearly optimal models, and achieve contextualized fairness. Our findings reveal that fairness can be enhanced without sacrificing model performance. By enhancing interpretability and fostering active engagement with domain experts based on a diverse set of nearly optimal models, FAIM promotes contextualized AI fairness, encouraging clinician involvement and multidisciplinary collaboration.

Enhancing model fairness without impairing model performance

Beyond merely enhancing fairness, it is crucial for bias mitigation methods to uphold satisfactory performance (e.g., AUC). A model’s performance is paramount; without it, the model’s utility in clinical decision-making is compromised.40 Some fairness methods, particularly post-process methods, often sacrifice model performance,41 raising concerns about their practicality. Our evaluations also reveal that these methods tend to reduce overall AUCs without reliably diminishing disparities (for example, equalized-odds post-process, as shown in Tables 2 and 3). This reduction in predictive performance is likely due to the randomness introduced in post-processing to achieve a specific fairness metric (e.g., equalized odds), which becomes even less effective with more subgroups, considering the intersectional effects of sensitive variables crucial in clinical settings.22,23 Pre-process methods such as reweighing also face challenges with an increasing number of subgroups, as they assume equal positive rates among all subgroups.16 Regularization-based methods may skew the optimization objective toward the majority class,25 while adversarial-learning-based methods can be unstable during training and sensitive to hyperparameters.42 Furthermore, bias mitigation methods often view fairness and performance as conflicting objectives, typically constraining one to optimize the other,17,18,19,20,21 focusing on the trade-off between performance and fairness and actually leaving a single model candidate at a fixed performance level. FAIM questions whether a single model can directly and accurately capture the fairness in each specific context and addresses that at a performance level, the range of model fairness can be wide, and these models may differ in coefficients or parameters (Figures 2 and S1). Anchoring the performance achieved by the model solely optimized for performance, FAIM investigates the diversity of model reliance on variables and its impact on fairness.

Statistical functionality, for example, odds ratios and their CIs, is also essential for elucidating model outputs and informing decision-making, as is commonly needed in clinical studies. Hence, it may be insufficient to merely provide binary predictions without details about why a model made a certain prediction or under what conditions the predictions hold, as is the case with most conventional bias mitigation methods. In the demonstration with the logistic regression model on both MIMIC-IV-ED and SGH-ED datasets, FAIM could output similar odds ratios and CIs for most less-sensitive variables, aligning with baseline models, with some adjustments to address fairness (Figure 5). These findings indicate that FAIM can provide a viable and fairness-aware alternative to the original machine learning models that are optimized for performance.

The importance of interpretability in fairness modeling

The interpretability in FAIM is multilayered. Firstly, its underlying principle is intuitive, which is to prioritize fairer models among a set of nearly optimal models to improve fairness. The model family to which the fairness-unaware model as well as all the nearly optimal models belong is inherently interpretable. In addition, the process of model selection is transparent and interactive, as shown in Figure 3 (described in the subsection model selection with contextualized fairness in the experimental procedures). Such an interactive process can assist in integrating data-driven evidence and clinical expertise, especially regarding sensitive variables that require contextualized consideration rather than a one-fits-all approach. Moreover, the variable importance analyses (e.g., using SHAP in our examples) can illustrate changes in variable contribution between fairness-unaware and fairness-aware models, for example, the diminished importance of the variable “pain description” on MIMIC-IV-ED data.

While many methods may produce bias-mitigated predictions, they may not always fundamentally address algorithmic bias or elucidate the modifications made to enhance fairness.25 For example, the post-process method FaiRS developed by Coston et al.43 is also grounded in near optimality. FaiRS thoughtfully adjusts binary predictions to promote group fairness, occasionally by altering a negative prediction to a positive one under the selective labels setting. FAIM further underscores the role of model interpretability in improving fairness and showcases that model interpretability can serve as a tool to disclose model information, verify the effects of bias mitigation strategies, and potentially enhance trust.

Keeping domain experts in the loop

Clinical evidence plays a crucial role in reducing model biases, especially when dealing with sensitive variables in the decision-making process. The results of FAIM in the MIMIC-IV-ED case resonate with literature that highlights racial and sex disparities in ED care outcomes in the US.36,37,38 Beyond excluding sensitive variables, FAIM also attenuated the variable of pain description, which may carry inherent sex biases.44,45

Model fairness can have geographical variations.23 As detailed in Table 4, the original models displayed race and sex disparities that were more pronounced in the US dataset compared to the Singapore dataset, and FAIM also showed varying effects in addressing these disparities. Additionally, FAIM addresses disparities across both race and sex in the US data while addressing primarily race in the Singapore data.

Advancing AI fairness in healthcare necessitates domain expertise to steer the process of bias mitigation effectively. FAIM generates and operates within a set of nearly optimal models that maintain performance comparable to the optimal model while exhibiting varying degrees of variable reliance (e.g., temperature), leading to various fairness profiles (Figures 2, 3, S1, and S2). Such diversity in fairness within the near optimality of performance allows for meaningful interactions between domain experts and algorithm developers to address specific fairness issues within the clinical context. Particularly, FAIM’s user-friendly graphical interface enables clinicians to influence model development with their insights directly. It also helps the validation of clinical judgment with data-driven evidence and encourages dialogue in the presence of divergences. For example, if including sensitive variables improves fairness, then this may suggest a need to investigate systematic biases or verify corresponding biological differences across subgroups, which can be embedded within involved less-sensitive variables. Sensitive variables may act as proxies for model adjustment, as in association studies, to better estimate the treatment effects. However, in prediction models, including such variables can directly influence predicted risk and associated actions, potentially raising ethical concerns. Through FAIM, clinicians can verify and reinforce this ethical consideration, potentially leading to more informed and ethically sound model adjustments. These insights on fairness notions, critical metrics, and clinical evidence related to sensitive variables can ultimately shape a more equitable and contextualized approach to patient care.

The scalability of FAIM and future works

FAIM advocates that bias mitigation should occur concurrently with the modeling process, instead of as a one-time step, given that all stages can inject bias into the final decision-making.11 Despite FAIM being an in-process-type method, it can synergize with pre-process strategies like reweighing to tackle data underrepresentation, which is a known source of data bias, particularly when subgroup patterns are pronounced.41 Additionally, methods that uncover data biases and abnormalities46,47 can be integrated into the framework of FAIM for data pre-processing.

In addition to its flexibility in module plug in, the framework of FAIM also offers adaptability to other machine learning models (e.g., neural networks) in the future. Such adaption requires additional work for different model types.31 The sampling distribution of coefficients may not always be applicable; however, it could be executed via empirical procedures, such as generating random masks for weight matrices for neural networks, or model-specific theoretical techniques, such as dynamic programming methods to enumerate the trees for decision trees.31 Catering to the model types, other explanation methods (e.g., Grad-CAM48 for deep learning models) can be used beyond SHAP to more accurately explain model behaviors to investigate the model changes when addressing fairness.

Although our study focused on separation-based metrics like equalized odds, equal opportunity, and BER equality, which are widely used in practice and relate directly to essential model performance, the FAIM framework can also integrate other widely used fairness definitions, for example, independence-based metrics14 such as statistical parity into the FRI composition. Independence-based metrics are useful when the outcome of interest lacks a reliable ground truth, rendering separation-based metrics unreliable.49 However, caution is advised in using independence-based metrics given the risk of misapplication, particularly in complex contexts of healthcare where biological differences and social biases often intertwine with each other.11,49,50 Moreover, despite current fairness metrics predominantly addressing binary outcomes,14 extending FAIM to handle other outcome types (e.g., time-to-event outcomes51) with customized fairness metrics is a promising direction for future work.

While FAIM was designed with a focus on clinical AI fairness, its versatility extends to other high-stakes domains, such as finance and criminal justice. See Note S1 for the summarized results using the COMPAS data,52 showing the benefits of FAIM when forcing the exclusion of sensitive variables from modeling. Notably, achieving AI fairness can present inherently distinct challenges across fields. The involvement of domain experts in the modeling process is indispensable, as their deep expertise can ensure that models are grounded in the nuanced realities of each domain. This further emphasizes the crucial role of keeping humans in the loop to navigate the complexities effectively.

Limitations

This study has several limitations. Firstly, we only demonstrated our method in two datasets related to emergency medicine. Further validation is needed in a wide spectrum of clinical applications. Additionally, the procedure of pinpointing nearly optimal models may benefit from further optimization. Lastly, there is room for improvement in the manner of obtaining CIs for the alternative odds ratios to more closely align with the functionality of the original models.

Conclusion

FAIM is an interpretable and interactive framework capable of generating fairness-aware models. These models yielded by FAIM can achieve a significant improvement in fairness without compromising the performance and functionality. The interpretability of this approach invites domain experts into the heart of the modeling process, fostering a richer multidisciplinary collaboration. This collaboration is crucial for crafting AI fairness to specific contexts, thereby aiding in fair and just clinical decision-making and ultimately advancing equitable AI in healthcare.

Experimental procedures

The FAIM framework

The FAIM framework consists of three modules to locate a “fairer” model in a set of nearly optimal models: nearly optimal model generation, model selection with contextualized fairness, and model explanation, as visualized in Figure 1. Let Y denote the outcome, Yˆ denote the predictions, XS = (Xs1,,Xsk) denote a set of k sensitive variables (e.g., sex and race), XS represent a feature subset of sensitive variables in XS, and XU=(XU1,,XUp) collectively denote p less-sensitive variables. Notably, XU are variables that do not directly contain sensitive information and cause bias, but as shown in the clinical examples, some seemingly non-sensitive variables may indirectly link to sensitive information and be affected after fairness enhancement. A model built with variables XU and XS is denoted by f(XU,XS). The loss function of this model is denoted as L(f(XU,XS),Y), along with the expected loss E[L(f(XU,XS),Y)].

Nearly optimal model generation: Composing IRSs with multiple sensitive variables

Conventional efforts to develop an optimal model typically focus on model performance, which is to fit the data and minimize the expected loss functions. Nevertheless, as one approaches the theoretically optimal model due to the data uncertainty, many models can fit the data with comparable loss, forming a set of nearly optimal models—an intriguing phenomenon known as the Rashomon effect in statistics.30,31,53 Within this Rashomon set of nearly optimal models, models can differ in their reliance on covariate information.32,54 This diversity leads to variations in fairness profiles, especially concerning sensitive variables, as displayed in Figures 2 and 3.

Let f(U,S) denote the optimal model that minimizes the expected loss in the model family F(U,S) that is built with less-sensitive variables XU and sensitive variables XS. The S-specific Rashomon set is defined as

R(U,S)(ϵ0,f(U,S),F(U,S))={fF(U,S)E[L(f,Y)](1+ϵ0)E[L(f(U,S),Y)]}

where near optimality is controlled by the small factor ϵ0>0. Particularly, for parametric models that can be fully represented by their coefficients β(U,S), i.e., a transparent model (for example, the family of logistic regression), the Rashomon set can be converted into a set of coefficients:

R(U,S)(ϵ0,β(U,S),Β(U,S))={βΒ(U,S)E[L(fβ,Y)](1+ϵ0)E[L(f(U,S),Y)]},

where Β(U,S) is the coefficient space of models in F(U,S) and β(U,S) is the coefficients of the optimal model f(U,S). For practical usage, the anchor of the Rashomon set—the expected loss—can also be replaced by a performance metric M (such as the AUC value) evaluated on the validation set, leading to

R(U,S)(ϵ0,β(U,S),Β(U,S))={βΒ(U,S)M(fβ,Y)(1ϵ0)M(f(U,S),Y)}.

To objectively evaluate the effects of sensitive variables on model performance and fairness, we consider different cases of variable selection for XS. The cases include no exclusion (i.e., X(U,S), the baseline case), complete exclusion (i.e., XU, under blindness), and all possible cases of partial exclusion. We defined the integral Rashomon set (IRS) with all the cases of S considered that can be expressed as

R(ϵ,β(U,·),Β(U,·))=SSβ(U,S)R(U,S)(ϵ0,β(U,S),Β(U,S))R(U,S)(ϵ0,β(U,S),Β(U,S)).

The case-specific near optimality determined by ϵ0 should be more stringent to guarantee the overall near optimality of the IRS determined by ϵ, i.e., ϵ>ϵ0>0 (see more details in supplemental experimental procedures S1). In previous studies, ϵ is often set at 5%.32,54

We employ the method of rejection sampling32,55 to identify the nearly optimal models, that is, to generate random samples of the coefficient vector and reject those with corresponding expected loss (or performance) out of near optimality. Specifically, to fully represent the case-specific Rashomon set R(U,S)(ϵ0,β(U,S),Β(U,S)), the i-th sample of the coefficient vector is generated from a multivariable normal distribution N(β(U,S),kiΣ(U,S)), where β(U,S) and ΣU,S are the coefficient vector and variance-covariance matrix of the optimal model f(U,S) based on less-sensitive variables XU and sensitive variables XS, respectively. Sampling guided by the characteristics of the optimal model enables the efficient generation of nearly optimal models and addresses model diversity within near optimality of performance. This approach can generate models with varied coefficients, as shown in Figure 2, leading to various fairness profiles while achieving nearly optimal performance as required by the reject sampling, which allows for the identification of fairer models. In addition, the control parameter of scope-width ki is drawn from a uniform distribution U(u1,u2) with tunable parameters u1 and u2 to adjust the scope of sampling.

Model selection with contextualized fairness

To comprehensively rank the nearly optimal models in the IRS, i.e., R(ϵ,β(U,·),Β(U,·)), and select the fairer one, we developed the FRI. Inspired by the radar chart for comparing items across multiple dimensions, the FRI is a holistic ranking measure that considers not only individual dimensions of fairness metrics (m1,m2,,mJ) but also their interdependencies, calculated as

FRI(fβ)=1j=1Jmj(fβ)mj+1(fβ),

where mJ+1=m1 to simplify notations. The highest-ranked model within R(ϵ,β(U,·),Β(U,·)) is chosen as the default fairness-aware model:

β˜=argmaxβR(ϵ,β(U,·),Β(U,·))FRI(fβ),

with a corresponding subset of sensitive variable(s) XS˜. The FAIM framework prioritizes the fairness notion of ensuring the equality of model discriminatory performance across subgroups (also named “separation-based” fairness), focusing on making the prediction Yˆ independent of sensitive variables XS when conditioned on the true outcome Y.14,22,25,56,57 This fairness notion encompasses various fairness metrics that essentially measure the gaps in performance (e.g., accuracy, AUC, sensitivity, specificity, etc.) among subgroups, with smaller gaps indicating better fairness.56 For example, equal opportunity22 emphasizes equal sensitivity (i.e., true positive rate) across subgroups, defined as the maximal discrepancy in sensitivity values across subgroups. Particularly, we used commonly used statistical metrics, including equalized odds,22 equal opportunity,22 and BER equality,14 as default options to compose the FRI (see more details in Table 1).

The process of model selection is designed to be interpretable and interactive. To facilitate such human-involved investigation, FAIM provides an interactive graphical interface (https://github.com/nliulab/FAIM) to display detailed information on each model upon hovering. As illustrated in Figure 3, which is based on SGH-ED data, Figure 3A visualized individual fairness metrics in each panel for the sampled nearly optimal models, with joint fairness ranked by the FRI. The top 1 model highlighted in gold is the default choice for the fairness-aware model. Nevertheless, users can assess alternative models with high rankings, such as the top 10 models, through a clinical lens, based on detailed data-driven evidence provided in Figures 3B and 3C. These high-ranked models may differ in how they handle sensitive variables yet have comparable model performance and fairness status. Clinical justification can jump in to support the final exclusion of sensitive variables. Therefore, the final fairness-aware model yielded by FAIM integrates domain knowledge with data-driven evidence to enhance model fairness without impairing performance.

As depicted in Figure 1, we investigate how variable importance shifts before and after fairness enhancement via variable importance analysis with SHAP. While we can directly compare the coefficients of the two logistic regression models, the value of coefficients is often not directly comparable between continuous and categorical variables.58 Therefore, we used SHAP as an alternative measure of variable importance to facilitate comparison between the fairness-unaware and fairness-aware models. The fairness-aware model fβ˜ shares a similar model architecture with the fairness-unaware fβ(U,S), except it may exclude certain sensitive variables that do not contribute to model performance and degraded fairness. The vector β˜ serves as a nearly optimal solution that minimally deviates from β(U,S) to execute fairness adjustments. In addition, fβ˜ holds the same functionality as f(U,S) in implementation. Take logistic regression for example: the coefficient β˜ can still be interpreted as the vector of log values of odds ratios, corresponding to the model fβ˜=11+eβ˜X(U,S˜) and logit loss L(fβ˜,Y). The standard errors for β˜ are estimated using Fisher’s information I(β˜), mirroring the method used for β(U,S) based on I(β(U,S)).

Data and study design

To illustrate the clinical application of the FAIM framework, we used two datasets from different populations. The first dataset was derived from the MIMIC-IV-ED,33 a publicly available database of ED admissions at the Beth Israel Deaconess Medical Center (BIDMC) in the US between 2011 and 2019. The second dataset, referred to as the SGH-ED database,34 was acquired from a tertiary hospital in Singapore, with data of approximately 1.8 million ED visits between 2008 and 2020.

Our predictive models for hospital admission (Y), informed by a previous study,59 utilized predictors including demographic data, vital signs, triage score, and health records (see more details in Table S4). For MIMIC-IV-ED data, we excluded patients with an age below 18,59 while for SGH-ED data, we excluded patients who were not Singapore residents or were aged below 21.54 In MIMIC-IV-ED data, we recategorized race into five groups—Asian, Black, Hispanic, White, and others—and binarized the ESI into low risk (3–5) and high risk (1–2).60 In SGH-ED data, we recategorized race into four groups, Chinese, Indian, Malay, and others, and recategorized the triage score—PACS—into three levels: P1, P2, and P3–4.61 Given the interplay of racism and sexism,23,24 we treated race and sex as sensitive variables in predicting hospital admissions.

Each dataset was randomly split into training (70%), validation (10%), and test (20%) sets. We utilized the training set for the generation of nearly optimal models, the validation set for model selection with contextualized fairness, and the test set for model evaluation. In addition, to illustrate variable importance using SHAP, we utilized subsamples of training and validation sets, respectively, as background and explanation data.

Statistical analysis

To compare the effects on fairness improvement from the baseline model, we included the well-validated pre-process method reweighing,16 the in-process methods reductions19 and Fair-GLM62 (regularization-based methods), and the post-process method equalized-odds post-processing22 for comparison. These methods of manipulating logistic regression to achieve fairness share a similar level of transparency with respect to the model setting. Additionally, we also included Adnet,25 a recent bias mitigation method that utilizes black-box models, particularly adversarial learning, to compare fairness achievement. See supplemental experimental procedures S2 for more implementation details. A conventionally trained logistic regression model served as the fairness-unaware baseline model. We evaluated model fairness with equalized odds, equal opportunity, and BER equality. We assessed model performance using the AUC, sensitivity, and specificity values, with 95% CIs reported. The sensitivity and specificity values were yielded based on optimal thresholds determined by Youden’s J statistics.63 The performance of two bias mitigation methods was deemed comparable if their AUC CIs overlapped. The data analysis and model building were performed using R v.4.0.2 (The R Foundation for Statistical Computing) and Python v.3.9.7.

Resource availability

Lead contact

Further information and requests for resources should be directed to and will be fulfilled by the lead contact, Nan Liu (liu.nan@duke-nus.edu.sg).

Materials availability

This research did not generate any materials.

Data and code availability

The MIMIC-IV-ED data are publicly available subject to the completion of ethics training and a signed data use agreement and are for research only. Due to ethical reasons and institutional guidelines, the SGH-ED data cannot be shared publicly. Data are available to researchers with some access restrictions applied upon request. Interested researchers may contact the SingHealth Health Services Research Center (hsr@singhealth.com.sg) for more details. The original codes have been deposited at Zenodo64 and are publicly available at GitHub, https://github.com/nliulab/FAIM.

Acknowledgments

We thank Prof. Bin Yu and Mr. Zachary Thomas Rewolinski for their suggestions on improving this paper. This work was supported by the Duke-NUS Signature Research Programme funded by the Ministry of Health, Singapore. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not reflect the views of the Ministry of Health. Y.N. was supported by the Khoo Postdoctoral Fellowship Award (project no. Duke-NUS-KPFA/2021/0051) from the Estate of Tan Sri Khoo Teck Puat.

Author contributions

M.L. and N.L. conceived and designed the study. M.L. and Y.N. wrote the code. M.L. performed the analyses and wrote the manuscript. M.L., Y.N., and N.L. made critical revisions to the manuscript. All authors interpreted the content, made revisions to the manuscript, and had final approval of the completed version. N.L. oversaw the project.

Declaration of interests

The authors declare no competing interests.

Declaration of generative AI and AI-assisted technologies in the writing process

During the preparation of this work, the authors used ChatGPT4.0 to improve the readability and language of the manuscript. After using this tool/service, the authors reviewed and edited the content as needed and take full responsibility for the content of the published article.

Published: September 12, 2024

Footnotes

Supplemental information can be found online at https://doi.org/10.1016/j.patter.2024.101059.

Supplemental information

Document S1. Figures S1–S4, Tables S1–S6, Note S1, and supplemental experimental procedures
mmc1.pdf (659.4KB, pdf)
Document S2. Article plus supplemental information
mmc2.pdf (4.1MB, pdf)

References

  • 1.Haug C.J., Drazen J.M. Artificial Intelligence and Machine Learning in Clinical Medicine, 2023. N. Engl. J. Med. 2023;388:1201–1208. doi: 10.1056/NEJMra2302038. [DOI] [PubMed] [Google Scholar]
  • 2.Bohr A., Memarzadeh K. Artificial Intelligence in Healthcare. Academic Press; 2020. The rise of artificial intelligence in healthcare applications; pp. 25–60. [Google Scholar]
  • 3.DeCamp M., Lindvall C. Mitigating bias in AI at the point of care. Science. 2023;381:150–152. doi: 10.1126/science.adh2713. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Mertens M. The Rowman & Littlefield Handbook of Bioethics. Rowman & Littlefield; 2022. Bias in Medicine. [Google Scholar]
  • 5.Brown A., Tomasev N., Freyberg J., Liu Y., Karthikesalingam A., Schrouff J. Detecting shortcut learning for fair medical AI using shortcut testing. Nat. Commun. 2023;14:4314. doi: 10.1038/s41467-023-39902-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Sjoding M.W., Dickson R.P., Iwashyna T.J., Gay S.E., Valley T.S. Racial Bias in Pulse Oximetry Measurement. N. Engl. J. Med. 2020;383:2477–2478. doi: 10.1056/NEJMc2029240. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Kozhimannil K.B., Trinacty C.M., Busch A.B., Huskamp H.A., Adams A.S. Racial and ethnic disparities in postpartum depression care among low-income women. Psychiatr. Serv. 2011;62:619–625. doi: 10.1176/appi.ps.62.6.619. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Panch T., Mattie H., Atun R. Artificial intelligence and algorithmic bias: implications for health systems. J. Glob. Health. 2019;9 doi: 10.7189/jogh.09.020318. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Fletcher R.R., Nakeshimana A., Olubeko O. Addressing Fairness, Bias, and Appropriate Use of Artificial Intelligence and Machine Learning in Global Health. Front. Artif. Intell. 2020;3 doi: 10.3389/frai.2020.561802. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Tsai T.C., Arik S., Jacobson B.H., Yoon J., Yoder N., Sava D., Mitchell M., Graham G., Pfister T. Algorithmic fairness in pandemic forecasting: lessons from COVID-19. NPJ Digit. Med. 2022;5:59. doi: 10.1038/s41746-022-00602-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Liu M., Ning Y., Teixayavong S., Mertens M., Xu J., Ting D.S.W., Cheng L.T.-E., Ong J.C.L., Teo Z.L., Tan T.F., et al. A translational perspective towards clinical AI fairness. NPJ Digit. Med. 2023;6:172. doi: 10.1038/s41746-023-00918-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Cirillo D., Catuara-Solarz S., Morey C., Guney E., Subirats L., Mellino S., Gigante A., Valencia A., Rementeria M.J., Chadha A.S., Mavridis N. Sex and gender differences and biases in artificial intelligence for biomedicine and healthcare. NPJ Digit. Med. 2020;3:81–81.11. doi: 10.1038/s41746-020-0288-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Gupta U., Ferber A.M., Dilkina B., Ver Steeg G. Controllable Guarantees for Fair Outcomes via Contrastive Information Estimation. Proc. AAAI Conf. Artif. Intell. 2021;35:7610–7619. [Google Scholar]
  • 14.Caton S., Haas C. Fairness in Machine Learning: A Survey. ACM Comput. Surv. 2024;56:1–38. [Google Scholar]
  • 15.Mehrabi N., Morstatter F., Saxena N., Lerman K., Galstyan A. A Survey on Bias and Fairness in Machine Learning. ACM Comput. Surv. 2021;54:1–35. [Google Scholar]
  • 16.Kamiran F., Calders T. Data preprocessing techniques for classification without discrimination. Knowl. Inf. Syst. 2012;33:1–33. [Google Scholar]
  • 17.Kamiran F., Calders T. Proceedings of Control and Communication 2009 2nd International Conference on Computer. 2009. Classifying without discriminating; pp. 1–6. [Google Scholar]
  • 18.Liu X.J., Li B., Bron E.E., Niessen W.J., Wolvius E.B., Roshchupkin G.V. Projection-Wise Disentangling for Fair and Interpretable Representation Learning: Application to 3D Facial Shape Analysis. Proceedings of International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) 2021;12905:814–823. [Google Scholar]
  • 19.Agarwal A., Beygelzimer A., Dudík M., Langford J., Wallach H. International Conference on Manchine Learning. PMLR; 2018. A reductions approach to fair classification; pp. 60–69. [Google Scholar]
  • 20.Madras D., Creager E., Pitassi T., Zemel R. Proceedings of the Conference on Fairness, Accountability, and Transparency. 2019. Fairness through Causal Awareness: Learning Causal Latent-Variable Models for Biased Data; pp. 349–358. [Google Scholar]
  • 21.Zhao Q., Adeli E., Pohl K.M. Training confounder-free deep learning models for medical applications. Nat. Commun. 2020;11:6010. doi: 10.1038/s41467-020-19784-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Hardt M., Price E., Price E., Srebro N. Equality of Opportunity in Supervised Learning. Proceedings of the 30th International Conference on Neural Information Processing Systems. 2016;29 [Google Scholar]
  • 23.Lett E., La Cava W.G. Translating intersectionality to fair machine learning in health sciences. Nat. Mach. Intell. 2023;5:476–479. doi: 10.1038/s42256-023-00651-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Mhasawade V., Zhao Y., Chunara R. Machine learning and algorithmic fairness in public and population health. Nat. Mach. Intell. 2021;3:659–666. [Google Scholar]
  • 25.Yang J., Soltan A.A.S., Eyre D.W., Yang Y., Clifton D.A. An adversarial training framework for mitigating algorithmic biases in clinical machine learning. NPJ Digit. Med. 2023;6:55. doi: 10.1038/s41746-023-00805-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Kamiran F., Karim A., Zhang X. 2012 IEEE 12th international conference on data mining. 2012. Decision Theory for Discrimination-Aware Classification; pp. 924–929. [Google Scholar]
  • 27.Black E., Raghavan M., Barocas S. Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency. Association for Computing Machinery; 2022. Model Multiplicity: Opportunities, Concerns, and Solutions; pp. 850–863. [Google Scholar]
  • 28.Chen Z., Zhang J.M., Sarro F., Harman M. A Comprehensive Empirical Study of Bias Mitigation Methods for Machine Learning Classifiers. ACM Trans. Software Eng. Methodol. 2023;32:1–30. [Google Scholar]
  • 29.Pfohl S.R., Foryciarz A., Shah N.H. An empirical characterization of fair machine learning for clinical risk prediction. J. Biomed. Inf. 2021;113 doi: 10.1016/j.jbi.2020.103621. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Breiman L. Statistical Modeling: The Two Cultures (with comments and a rejoinder by the author) Stat. Sci. 2001;16:199–231. [Google Scholar]
  • 31.Dong J., Rudin C. Exploring the cloud of variable importance for the set of all good models. Nat. Mach. Intell. 2020;2:810–824. [Google Scholar]
  • 32.Ning Y., Ong M.E.H., Chakraborty B., Goldstein B.A., Ting D.S.W., Vaughan R., Liu N. Shapley variable importance cloud for interpretable machine learning. Patterns. 2022;3 doi: 10.1016/j.patter.2022.100452. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Johnson A., Bulgarelli L., Pollard T., Celi L.A., Mark R., Horng S. 2023. MIMIC-IV-ED. PhysioNet. [DOI] [Google Scholar]
  • 34.Liu N., Xie F., Siddiqui F.J., Ho A.F.W., Chakraborty B., Nadarajan G.D., Tan K.B.K., Ong M.E.H. Leveraging Large-Scale Electronic Health Records and Interpretable Machine Learning for Clinical Decision Making at the Emergency Department: Protocol for System Development and Validation. JMIR Res. Protoc. 2022;11 doi: 10.2196/34201. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Rudin C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019;1:206–215. doi: 10.1038/s42256-019-0048-x. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Zhang X., Carabello M., Hill T., Bell S.A., Stephenson R., Mahajan P. Trends of Racial/Ethnic Differences in Emergency Department Care Outcomes Among Adults in the United States From 2005 to 2016. Front. Med. 2020;7:300. doi: 10.3389/fmed.2020.00300. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Kam-lun E.H., Nelson A. Gender disparity in paediatric hospital admissions. Ann. Acad. Med. Singapore. 2006;35:882–888. [PubMed] [Google Scholar]
  • 38.Baibergenova A., Thabane L., Akhtar-Danesh N., Levine M., Gafni A., Leeb K. Sex differences in hospital admissions from emergency departments in asthmatic adults: a population-based study. Ann. Allergy Asthma Immunol. 2006;96:666–672. doi: 10.1016/S1081-1206(10)61063-0. [DOI] [PubMed] [Google Scholar]
  • 39.Lundberg S., Lee S.-I. A Unified Approach to Interpreting Model Predictions. arXiv. 2017 doi: 10.48550/arXiv.1705.07874. Preprint at. [DOI] [Google Scholar]
  • 40.Rajkomar A., Oren E., Chen K., Dai A.M., Hajaj N., Hardt M., Liu P.J., Liu X., Marcus J., Sun M., et al. Scalable and accurate deep learning with electronic health records. NPJ Digit. Med. 2018;1 doi: 10.1038/s41746-018-0029-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Petersen E., Holm S., Ganz M., Feragen A. The path toward equal performance in medical machine learning. Patterns. 2023;4 doi: 10.1016/j.patter.2023.100790. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Zhang B.H., Lemoine B., Mitchell M. Proceedings of the 2018 AAAI/ACM Conference on AI, Ethics, and Society. Association for Computing Machinery; 2018. Mitigating Unwanted Biases with Adversarial Learning; pp. 335–340. [Google Scholar]
  • 43.Coston A., Rambachan A., Chouldechova A. International Conference on Machine Learning (PMLR) 2021. Characterizing Fairness Over the Set of Good Models Under Selective Labels; pp. 2144–2155. [Google Scholar]
  • 44.Patel R., Biros M.H., Moore J., Miner J.R. Gender differences in patient-described pain, stress, and anxiety among patients undergoing treatment for painful conditions in the emergency department. Acad. Emerg. Med. 2014;21:1478–1484. doi: 10.1111/acem.12543. [DOI] [PubMed] [Google Scholar]
  • 45.Marco C.A., Kanitz W., Jolly M. Pain Scores among Emergency Department (ED) Patients: Comparison by ED Diagnosis. J. Emerg. Med. 2013;44:46–52. doi: 10.1016/j.jemermed.2012.05.002. [DOI] [PubMed] [Google Scholar]
  • 46.Ghorbani A., Zou J. International Conference on Machine Learning. PMLR; 2019. Data Shapley: Equitable Valuation of Data for Machine Learning; pp. 2242–2251. [Google Scholar]
  • 47.Jiang H., Nachum O. International conference on artificial intelligence and statistics. PMLR); 2020. Identifying and Correcting Label Bias in Machine Learning; pp. 702–712. [Google Scholar]
  • 48.Selvaraju R.R., Cogswell M., Das A., Vedantam R., Parikh D., Batra D. Proceedings of the IEEE international conference on computer vision. 2017. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization; pp. 618–626. [Google Scholar]
  • 49.Weerts H., Dudík M., Edgar R., Jalali A., Lutz R., Madaio M. Fairlearn: Assessing and Improving Fairness of AI Systems. J. Mach. Learn. Res. 2023;24:1–8. [Google Scholar]
  • 50.Watkins E.A., McKenna M., Chen J. The four-fifths rule is not disparate impact: a woeful tale of epistemic trespassing in algorithmic fairness. arXiv. 2022 https://arxiv.org/abs/2202.09519 Preprint at. [Google Scholar]
  • 51.Rahman M.M., Purushotham S. Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining. Association for Computing Machinery; 2022. Fair and Interpretable Models for Survival Analysis; pp. 1452–1462. [Google Scholar]
  • 52.Larson J., Mattu S., Kirchner L., Angwin J. 2016. How We Analyzed the Compas Recidivism Algorithm.https://www.propublica.org/article/how-we-analyzed-the-compas-recidivism-algorithm [Google Scholar]
  • 53.Rudin C., Zhong C., Semenova L., Seltzer M., Parr R., Liu J., Katta S., Donnelly J., Chen H., Boner Z. Amazing Things Come From Having Many Good Models. arXiv. 2024 doi: 10.48550/arXiv.2407.04846. Preprint at. [DOI] [Google Scholar]
  • 54.Ning Y., Li S., Ong M.E.H., Xie F., Chakraborty B., Ting D.S.W., Liu N. A novel interpretable machine learning system to generate clinical risk scores: An application for predicting early mortality or unplanned readmission in a retrospective cohort study. PLOS Digit. Health. 2022;1 doi: 10.1371/journal.pdig.0000062. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Yu B., Kumbier K. Veridical data science. Proc. Natl. Acad. Sci. USA. 2020;117:3920–3929. doi: 10.1073/pnas.1901326117. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Xu J., Xiao Y., Wang W.H., Ning Y., Shenkman E.A., Bian J., Wang F. Algorithmic fairness in computational medicine. EBioMedicine. 2022;84 doi: 10.1016/j.ebiom.2022.104250. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Seyyed-Kalantari L., Zhang H., McDermott M.B.A., Chen I.Y., Ghassemi M. Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations. Nat. Med. 2021;27:2176–2182. doi: 10.1038/s41591-021-01595-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Ning Y., Li S., Ng Y.Y., Chia M.Y.C., Gan H.N., Tiah L., Mao D.R., Ng W.M., Leong B.S.-H., Doctor N., et al. Variable importance analysis with interpretable machine learning for fair risk prediction. PLOS Digit. Health. 2024;3 doi: 10.1371/journal.pdig.0000542. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Xie F., Zhou J., Lee J.W., Tan M., Li S., Rajnthern L.S., Chee M.L., Chakraborty B., Wong A.-K.I., Dagan A., et al. Benchmarking emergency department prediction models with machine learning and public electronic health records. Sci. Data. 2022;9:658. doi: 10.1038/s41597-022-01782-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Emergency Severity Index (ESI): A triage tool for emergency departments. https://www.ahrq.gov/patient-safety/settings/emergency-dept/esi.html.
  • 61.Fong R.Y., Glen W.S.S., Mohamed Jamil A.K., Tam W.W.S., Kowitlawakul Y. Comparison of the Emergency Severity Index versus the Patient Acuity Category Scale in an emergency setting. Int. Emerg. Nurs. 2018;41:13–18. doi: 10.1016/j.ienj.2018.05.001. [DOI] [PubMed] [Google Scholar]
  • 62.Do H., Putzel P., Martin A., Smyth P., Zhong J. Vol. 162. PMLR; 2022. Fair Generalized Linear Models with a Convex Penalty; pp. 5286–5308. (Proc. Mach. Learn. Res.). [PMC free article] [PubMed] [Google Scholar]
  • 63.Youden W.J. Index for rating diagnostic tests. Cancer. 1950;3:32–35. doi: 10.1002/1097-0142(1950)3:1<32::aid-cncr2820030106>3.0.co;2-3. [DOI] [PubMed] [Google Scholar]
  • 64.Liu M., Liu N. Zenodo; 2024. Source Code of “Fairness-Aware Interpretable Modelling (FAIM) for Trustworthy Machine Learning in Healthcare”. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Document S1. Figures S1–S4, Tables S1–S6, Note S1, and supplemental experimental procedures
mmc1.pdf (659.4KB, pdf)
Document S2. Article plus supplemental information
mmc2.pdf (4.1MB, pdf)

Data Availability Statement

The MIMIC-IV-ED data are publicly available subject to the completion of ethics training and a signed data use agreement and are for research only. Due to ethical reasons and institutional guidelines, the SGH-ED data cannot be shared publicly. Data are available to researchers with some access restrictions applied upon request. Interested researchers may contact the SingHealth Health Services Research Center (hsr@singhealth.com.sg) for more details. The original codes have been deposited at Zenodo64 and are publicly available at GitHub, https://github.com/nliulab/FAIM.


Articles from Patterns are provided here courtesy of Elsevier

RESOURCES