Risk assessment algorithms used in criminal justice settings are often said to introduce “bias”. But such charges can conflate an algorithm’s performance with bias in the data used to train the algorithm and with bias in the actions undertaken with an algorithm’s output. In this paper, algorithms themselves are the focus. Tradeoffs between different kinds of fairness and between fairness and accuracy are illustrated using an algorithmic application to juvenile justice data. Given potential bias in training data, can risk assessment algorithms improve fairness, and if so, with what consequences for accuracy? Although statisticians and computer scientists can documents the tradeoffs, they cannot provide technical solutions that satisfy all fairness and accuracy objectives. In the end, it falls to stakeholders to do the required balancing using legal and legislative procedures, just as it always has (p.1).
Accuracy and Fairness for Juvenile Justice Risks Assessments (2017)
Notice about external resources
These links are being provided as a convenience and for informational purposes only; they do not constitute an endorsement or an approval by the National Institute of Corrections (NIC) of any of the products, services or opinions of the corporation or organization or individual. NIC bears no responsibility for the accuracy, legality or content of the external site or for that of subsequent links. Contact the external site for answers to questions regarding its content.