This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

proofread

Analysis of court transcripts reveals biased jury selection

jury
Credit: Pixabay/CC0 Public Domain

Cornell researchers have shown that data science and artificial intelligence tools can successfully identify when prosecutors question potential jurors differently, in an effort to prevent women and Black people from serving on juries.

In a first-of-its-kind study, researchers used processing (NLP) tools to analyze transcripts of the jury selection process. They found multiple quantifiable differences in how prosecutors questioned Black and white members of the jury pool. Once validated, this technology could provide evidence for appeals cases and be used in real time during jury selection to ensure more diverse juries.

The study, "Quantifying Disparate Questioning of Black and White Jurors in Capital Jury Selection," was published July 14 in the Journal of Empirical Legal Studies. First author is Anna Effenberger.

Striking jurors on the basis of race or gender has been illegal since the Supreme Court's landmark Batson vs. Kentucky case in 1986, but this type of discrimination still occurs, said study co-author John Blume, the Samuel F. Leibowitz Professor of Trial Techniques at Cornell Law School and director of theCornell Death Penalty Project.

"One of the things the courts have looked at is whether the questions Black and white jurors differently," Blume said. "NLP software allows you to do that on a much more sophisticated level, looking at not just at the number, but the way in which the questions are put together."

Under the assumption that Black and female jurors will be more sympathetic to a defendant—especially a Black one—prosecutors will sometimes press them to reveal disqualifying information. A common tactic in capital cases is to provide an especially gruesome description of the execution process and then ask if the person would be willing to sentence the defendant to death. If the answer is no, that person is struck from the jury pool.

To see if NLP software could identify this and other signs of disparate questioning, Blume collaborated with Martin Wells, the Charles A. Alexander Professor of Statistical Sciences in the Cornell Ann. S Bowers College of Computing and Information Science, and Effenberger to analyze transcripts from 17 capital cases in South Carolina. Their dataset included more than 26,000 questions that judges, defense attorneys and the prosecution asked potential jurors.

The researchers looked not only at the number of questions asked of Black, white, male and female potential jurors, but also the topics covered, each question's complexity and the parts of speech used.

"We consistently found racial differences in a number of these measures," Wells said. "When we do job interviews, we usually have a list of questions, and we want to ask everyone the same question, and here that's not the case."

The analysis showed significant differences in the length, complexity and sentiment of the questions prosecutors asked of Black potential jurors compared to white ones, indicating they were likely attempting to shape their responses. The questions asked by judges and the defense showed no such .

The study also found evidence that prosecutors had attempted to disqualify Black individuals by using their views on the death penalty. Prosecutors asked Black potential jurors—especially those who were ultimately excused from serving—more explicit and graphic questions about execution methods compared to white potential jurors.

In six of the 17 cases analyzed in the study, a judge had later ruled that the prosecutor illegally removed potential jurors on the basis of race. By looking at the combined NLP analyses for each case, the researchers could successfully distinguish between cases that violated Batson vs. Kentucky, and ones that hadn't.

The researchers said the findings prove that NLP tools can successfully identify biased jury selection. Now, they hope to see similar studies performed on larger datasets with more diverse types of cases.

Once the validity of this method is established, "this could be done during jury selection almost in real time," Wells said.

Whether used to monitor selection or to provide evidence for an appeal, this software could be a powerful tool to diversify juries—especially for defendants who are potentially facing the death penalty.

More information: Anna Effenberger et al, Quantifying disparate questioning of Black and White jurors in capital jury selection, Journal of Empirical Legal Studies (2023). DOI: 10.1111/jels.12357

Provided by Cornell University

Citation: Analysis of court transcripts reveals biased jury selection (2023, July 28) retrieved 27 April 2024 from https://phys.org/news/2023-07-analysis-court-transcripts-reveals-biased.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Pretrial publicity hinders prosecutors' ability to prove guilt

45 shares

Feedback to editors