Healthcare-specific tools for bias analysis
There are three common categories of metrics for determining whether a model is considered “fair”: Group Fairness, which compares the statistical similarities of predictions relative to known and discrete protected groupings; Similarity-Based Measures, which evaluate predictions without those discrete protected groups; and Causal Reasoning measures, which evaluate fairness through the use of causal models.
Category | Metric | Definition | Weakness | References |
---|---|---|---|---|
Group Fairness | Demographic Parity | A model has Demographic Parity if the predicted positive rates (selection rates) are approximately the same for all protected attribute groups: Harms Addressed: Allocative |
Historical biases present in the data are not addressed and may still bias the model. | Zafar et al (2017) |
Equalized Odds | Odds are equalized if P(+) is approximately the same for all protected attribute groups. Equal Opportunity is a special case of equalized odds specifying that P(+ | y = 1) is approximately the same across groups. Harms Addressed: Allocative, Representational |
Historical biases present in the data are not addressed and may still bias the model. | Hardt et al (2016) | |
Predictive Parity | This parity exists where the Positive Predictive Value is approximately the same for all protected attribute groups. Harms Addressed: Allocative, Representational |
Historical biases present in the data are not addressed and may still bias the model. | Zafar et al (2017) | |
Similarity-Based Measures | Individual Fairness | Individual fairness exists if “similar” individuals (ignoring the protected attribute) are likely to have similar predictions. Harms Addressed: Representational |
The appropriate metric for similarity may be ambiguous. | Dwork (2012), Zemel (2013), Kim et al (2018) |
Entropy-Based Indices | Measures of entropy, particularly existing inequality indices from the field of economics, are applied to evaluate either individuals or groups Harms Addressed: Representational |
Speicher (2018) | ||
Unawareness | A model is unaware if the protected attribute is not used. Harms Addressed: Allocative, Representational |
Removal of a protected attribute may be ineffectual due to the presence of proxy features highly correlated with the protected attribute. | Zemel et al (2013), Barocas and Selbst (2016) | |
Causal Reasoning | Counterfactual Fairness * | Counterfactual fairness exists where counterfactual replacement of the protected attribute does not significantly alter predictive performance. This counterfactual change must be propogated to correlated variables. Harms Addressed: Allocative, Representational |
It may be intractable to develop a counterfactual model. | Russell et al (2017) |
Metric | Criteria | Definition | Description |
---|---|---|---|
Demographic Parity | Statistical Independence | Sensitive attributes (A) are statistically independent of the prediction result (R) | |
Equalized Odds | Statistical Separation | Sensitive attributes (A) are statistically independent of the prediction result (R) given the ground truth (Y) | |
Predictive Parity | Statistical Sufficiency | Sensitive attributes (A) are statistically independent of the ground truth (Y) given the prediction (R) |
From: Verma & Rubin, 2018
Name | Definition | About | Aliases | |
---|---|---|---|---|
Demographic Parity | Predictions must be statistically independent from the sensitive attributes. Subjects in all groups should have equal probability of being assigned to the positive class. Note: may fail if the distribution of the ground truth justifiably differs among groups Criteria: Statistical Independence |
Statistical Parity, Equal Acceptance Rate, Benchmarking | ||
Conditional Statistical Parity | Subjects in all groups should have equal probability of being assigned to the positive class conditional upon legitimate factors (L). Criteria: Statistical Separation |
|||
False positive error rate (FPR) balance | Equal probabilities for subjects in the negative class to have positive predictions. Mathematically equivalent to equal TNR: Criteria: Statistical Separation |
Predictive Equality | ||
False negative error rate (FNR) balance | Equal probabilities for subjects in the positive class to have negative predictions. Mathematically equivalent to equal TPR: Criteria: Statistical Separation |
Equal Opportunity | ||
Equalized Odds | Equal TPR and equal FPR. Mathematically equivalent to the conjunction of FPR balance and FNR balance Criteria: Statistical Separation |
Disparate mistreatment, Conditional procedure accuracy equality | ||
Predictive Parity | All groups have equal PPV (probability that a subject with a positive prediction actually belongs to the positive class. Mathematically equivalent to equal False Discovery Rate (FDR): Criteria: Statistical Sufficiency |
Outcome Test | ||
Conditional use accuracy equality | Criteria: Statistical Sufficiency | |||
Overall Accuracy Equity | Use when True Negatives are as desirable as True Positives | |||
Treatment Equality | Groups have equal ratios of False Negative Rates to False Positive Rates | |||
Calibration | For a predicted probability score S, both groups should have equal probability of belonging to the positive class Criteria: Statistical Sufficiency |
Test-fairness, matching conditional frequencies | ||
Well-calibration | For a predicted probability score S, both groups should have equal probability of belonging to the positive class, and this probability is equal to S Criteria: Statistical Sufficiency |
|||
Balance for positive class | Subjects in the positive class for all groups have equal average predicted probability score S Criteria: Statistical Separation |
|||
Balance for negative class | Subjects in the negative class for all groups have equal average predicted probability score S Criteria: Statistical Separation |
|||
Causal discrimination | Same classification produced for any two subjects with the exact same attributes | |||
Fairness through unawareness | No sensitive attributes are explicitly used in the decision-making process Criteria: Unawareness |
|||
Fairness through awareness (Individual Fairness) | for a set of applicants V , a distance metric between applicants k : V Å~V → R, a mapping from a set of applicants to probability distributions over outcomes M : V → δA, and a distance D metric between distribution of outputs, fairness is achieved iff |
Similar individuals (as defined by some distance metric) should have similar classification | Individual Fairness | |
Counterfactual fairness | A causal graph is counterfactually fair if the predicted outcome d in the graph does not depend on a descendant of the protected attribute G. | |||
Group Measure Type | Examples | “Fair” Range |
---|---|---|
Statistical Ratio | Disparate Impact Ratio, Equalized Odds Ratio | 0.8 <= “Fair” <= 1.2 |
Statistical Difference (Binary Classification) | Equalized Odds Difference, Predictive Parity Difference | -0.1 <= “Fair” <= 0.1 |
Statistical Difference (Regression) | MAE Difference, Mean Prediction Difference | Problem Specific |
Metric | Measure | Equation | Interpretation |
---|---|---|---|
Group Fairness Measures - Binary Classification | Selection Rate | - | |
Demographic (Statistical) Parity Difference | (-) favors privileged group (+) favors unprivileged group |
||
Disparate Impact Ratio (Demographic Parity Ratio) | < 1 favors privileged group > 1 favors unprivileged group |
||
Positive Rate Difference | (-) favors privileged group (+) favors unprivileged group |
||
Average Odds Difference | (-) favors privileged group (+) favors unprivileged group |
||
Average Odds Error | (-) favors privileged group (+) favors unprivileged group |
||
Equal Opportunity Difference | (-) favors privileged group (+) favors unprivileged group |
||
Equal Odds Difference | (-) favors privileged group (+) favors unprivileged group |
||
Equal Odds Ratio | < 1 favors privileged group > 1 favors unprivileged group |
||
Group Fairness Measures - Regression | Mean Prediction Ratio | < 1 favors privileged group > 1 favors unprivileged group |
|
Mean Prediction Difference | (-) favors privileged group (+) favors unprivileged group |
||
MAE Ratio | < 1 favors privileged group > 1 favors unprivileged group |
||
MAE Difference | (-) favors privileged group (+) favors unprivileged group |
||
Individual Fairness Measures | Consistency Score | 1 is consistent 0 is inconsistent |
|
Generalized Entropy Index | ![]() |
- | |
Generalized Entropy Error | - | ||
Between-Group Generalized Entropy Error | 0 is fair (+) is unfair |
Agarwal, A., Beygelzimer, A., Dudík, M., Langford, J., & Wallach, H. (2018). A reductions approach to fair classification. In International Conference on Machine Learning (pp. 60-69). PMLR. Available through arXiv preprint:1803.02453.
Barocas, S., & Selbst AD (201). Big data’s disparate impact. California Law Review, 104, 671. Retrieved from https://www.cs.yale.edu/homes/jf/BarocasDisparateImpact.pdf
Dwork, C., Hardt, M., Pitassi, T., Reingold, O., & Zemel, R. (2012, January). Fairness through awareness. In Proceedings of the 3rd innovations in theoretical computer science conference (pp. 214-226). Retrieved from https://arxiv.org/pdf/1104.3913.pdf
Hardt, M., Price, E., & Srebro, N. (2016). Equality of opportunity in supervised learning. In Advances in neural information processing systems (pp. 3315-3323). Retrieved from http://papers.nips.cc/paper/6374-equality-of-opportunity-in-supervised-learning.pdf
Kim, M., Reingol, O., & Rothblum, G. (2018). Fairness through computationally-bounded awareness. In Advances in Neural Information Processing Systems pp. 4842-4852). Retrieved from https://arxiv.org/pdf/1803.03239.pdf
Russell, C., Kusner, M.J., Loftus, J., & Silva, R. (2017). When worlds collide: integrating different counterfactual assumptions in fairness. In Advances in Neural Information Processing Systems (pp. 6414-6423). Retrieved from https://papers.nips.cc/paper/7220-when-worlds-collide-integrating-different-counterfactual-assumptions-in-fairness.pdf
Verma, S., & Rubin, J. (2018, May). Fairness definitions explained. In 2018 ieee/acm international workshop on software fairness (fairware) (pp. 1-7). IEEE.
Zemel, R., Wu, Y., Swersky, K., Pitassi, T., & Dwork, C. (2013, February). Learning fair representations. International Conference on Machine Learning (pp. 325-333). Retrieved from http://proceedings.mlr.press/v28/zemel13.pdf
Zafar, M.B., Valera, I., Gomez Rodriguez, M., & Gummadi, K.P. (2017, April). Fairness beyond disparate treatment & disparate impact: Learning classification without disparate mistreatment. In Proceedings of the 26th international conference on world wide web (pp. 1171-1180). https://arxiv.org/pdf/1610.08452.pdf