Study: Tools to assess crime risk for young cohorts are likely to fail over time if they ignore social change
Risk assessment instruments (RAIs) are widely used to inform high-stakes decision-making in the criminal justice system and other areas, such as health care and child welfare. These tools typically assume a relation between predictors and outcomes that does not vary with time. But because societies change, this assumption may not hold in all settings, generating what a new study calls cohort bias—a bias resulting from cohort-wide influences not experienced by past or future cohorts.
The study, by researchers at Carnegie Mellon University (CMU), Harvard, and University of Pennsylvania, appears in PNAS. The researchers analyzed criminal histories of individuals in Chicago over a 25-year period, 1995-2020. They found that regardless of predictors, a machine learning tool predicting the likelihood of arrest between ages 17 and 24 for older cohorts born in the 1980s substantially overpredicted the likelihood of arrest for younger cohorts born in the mid-1990s. This suggests that RAIs are likely to fail over time for more recent cohorts if social change dynamics are ignored.
"Our aim is to improve the science and use of RAIs in high-stakes contexts," says Daniel Nagin, professor of public policy and statistics at CMU's Heinz College, who goes on to point out, "Cohort bias can generate inequality in criminal justice that is distinct from racial bias."
In predicting criminal behavior and justice system involvement, RAIs typically predict risk using a combination of features that measure individual characteristics, family background, and prior criminal history.
Past research has identified early-life, psychosocial, and neighborhood predictors of criminal involvement in adolescence and beyond. These include family instability and poverty, low self-control and growing up in poor neighborhoods. While the association between many of these features and later criminal involvement is well established, whether the predictive strength of these features is constant across historical periods is uncertain.
In this study, researchers examined whether an RAI trained on individual-level features of an older-age cohort accurately predicted the likelihood of criminal justice involvement of a younger cohort. Specifically, they studied the performance of models trained to predict arrests in early adulthood across multiple birth cohorts born between 1979 and 1995.
Their data came from the longitudinal Project on Human Development in Chicago Neighborhoods and included more than 1,000 individuals from four age cohorts. Researchers also examined criminal records from the state of Illinois between 1995 and 2020. Among the study's findings:
- Cohort bias was significant: RAIs trained on an older cohort overpredicted the probability of arrest of the younger cohort by up to 89%.
- Cohort bias was substantial within racial-ethnic groups (whites and others, Latinos, and Blacks): This establishes cohort bias as an underappreciated mechanism generating inequality in criminal justice that is distinct from racial bias.
- Cohort bias persisted: It lasted even when measures of arrest from immediately before the ages for which the study predicted arrest were included as predictors and even when analysis was limited to high-risk participants.
As these findings suggest, individuals' future behaviors are not only a function of their stable traits, earlier life circumstances, prior behaviors, and age, but also of ongoing social changes that affect all members of a birth cohort. While the field understands that an algorithm's performance can degrade over time, the implications of that ongoing change are typically not recognized in RAIs or in the general conceptualization of future risk.
In the context of high-stakes decisions, RAIs have the potential to do harm if they are not well calibrated. While human judgment is also prone to biases, an RAI can make millions more judgments, amplifying its impact.
"Social science research and policy increasingly rely on predictive RAIs, including those that use machine learning algorithms," explains Erika Montana, a Ph.D. student in machine learning and public policy at CMU's Heinz College, who co-authored the study and took the lead role in conducting the analyses. "Our findings show that the relations between risk factors and future arrest are not stable over time. As a result, prediction models that rely on these risk factors are prone to systematic and substantial error."
More information: Erika Montana et al, Cohort bias in predictive risk assessments of future criminal justice system involvement, Proceedings of the National Academy of Sciences (2023). DOI: 10.1073/pnas.2301990120
Journal information: Proceedings of the National Academy of Sciences
Provided by Carnegie Mellon University's Heinz College