Building Fair Machine Learning Models: Using Big Data to Explore Inequities in Risk Assessment at the Centre for Addiction and Mental Health (CAMH)

Building Fair Machine Learning Models: Using Big Data to Explore Inequities in Risk Assessment at the Centre for Addiction and Mental Health (CAMH)

Big Data and advances in machine learning (ML) provide opportunities to improve risk assessments in mental health. Risk assessments can help manage uncertainty and account for past misfortunes. While few ML-based risk assessment tools are currently in use, it is easy to anticipate a future where such tools are used to help predict risk of inpatient violence, suicide, or psychiatric readmission. However, there is growing concern that ML tools amplify existing inequities, such as racial bias, often because they are trained on biased datasets. Few studies examine how risk assessments in mental health intersect with racism and other determinants of health, such as class, housing, and immigration status. Also, we know little about the experience of risk assessments from the patient’s perspective. Serious questions remain about how to account for biases and build fair models.

Additional info

Our project aims to integrate unstructured and structured risk assessment data from CAMH electronic health records (EHRs), identify various biases, and use ML to redress these biases and build fair models. As a step toward achieving this larger aim, our aim for this study is to provide the critical context for future applications of ML-based risk assessment tools.

Project team

  • Dr Daniel Buchman (Co-PI)
  • Dr Sean Hill (Co-PI)
  • Zoe Findlay
  • Dr Katrina Hui
  • Dr Marta Maslej
  • Darla Reslan
  • Dr Laura Sikstrom
  • Yifan Wang
  • Dr Juveria Zaheer

Funders and partners

  • Dalla Lana School of Public Health Data Science Interdisciplinary Research Seed Funding