Algorithmic fairness in precision psychiatry: analysis of prediction models in individuals at clinical high risk for psychosis

Derya Şahin*, Lana Kambeitz-Ilankovic, Stephen Wood, Dominic Dwyer, Rachel Upthegrove, Raimo Salokangas, Stefan Borgwardt, Paolo Brambilla, Eva Meisenzahl, Stephan Ruhrmann, Frauke Schultze-Lutter, Rebekka Lencer, Alessandro Bertolino, Christos Pantelis, Nikolaos Koutsouleris, Joseph Kambeitz

*Corresponding author for this work

Abstract

Background Computational models offer promising potential for personalised treatment of psychiatric diseases. For their clinical deployment, fairness must be evaluated alongside accuracy. Fairness requires predictive models to not unfairly disadvantage specific demographic groups. Failure to assess model fairness prior to use risks perpetuating healthcare inequalities. Despite its importance, empirical investigation of fairness in predictive models for psychiatry remains scarce. Aims To evaluate fairness in prediction models for development of psychosis and functional outcome. Method Using data from the PRONIA study, we examined fairness in 13 published models for prediction of transition to psychosis (n = 11) and functional outcome (n = 2) in people at clinical high risk for psychosis or with recent-onset depression. Using accuracy equality, predictive parity, false-positive error rate balance and false-negative error rate balance, we evaluated relevant fairness aspects for the demographic attributes 'gender' and 'educational attainment' and compared them with the fairness of clinicians' judgements. Results Our findings indicate systematic bias towards assigning less favourable outcomes to individuals with lower educational attainment in both prediction models and clinicians' judgements, resulting in higher false-positive rates in 7 of 11 models for transition to psychosis. Interestingly, the bias patterns observed in algorithmic predictions were not significantly more pronounced than those in clinicians' predictions. Conclusions Educational bias was present in algorithmic and clinicians' predictions, assuming more favourable outcomes for individuals with higher educational level (years of education). This bias might lead to increased stigma and psychosocial burden in patients with lower educational attainment and suboptimal psychosis prevention in those with higher educational attainment.

Original languageEnglish
JournalBritish Journal of Psychiatry
Volume224
Issue number2
Pages (from-to)55-65
Number of pages11
ISSN0007-1250
DOIs
Publication statusPublished - 08.02.2024

Research Areas and Centers

  • Academic Focus: Center for Brain, Behavior and Metabolism (CBBM)

DFG Research Classification Scheme

  • 2.23-10 Clinical Psychiatry, Psychotherapy, Child and Adolescent Psychiatry

Fingerprint

Dive into the research topics of 'Algorithmic fairness in precision psychiatry: analysis of prediction models in individuals at clinical high risk for psychosis'. Together they form a unique fingerprint.

Cite this