Shine a light on tech’s hidden impacts. Triple your donation.
Skip navigation

Hello World

Dropout Risk System Under Scrutiny After The Markup Report

The Center for Democracy & Technology makes a case for civil rights protections

Watercolor illustration of a snaking line of students wearing graduation caps. Ten of the students are pink and there’s a bracket above them with the text “Less than 78.5% predicted chance of graduating.”
In our investigation, we found that Wisconsin uses a computer model to predict how likely middle school students are to graduate from high school on time. Gabriel Hongsdusit

Hi there,

It’s me, Tara García Mathewson, back in the Hello World rotation to get you thinking about civil rights in education. 

Earlier this week, the Center for Democracy & Technology (CDT) put out a report arguing that decades-old civil rights laws should be protecting students from discrimination based on their race—but they’re not. Reporting from The Markup was at the heart of one of its examples.

Did you see the story Todd Feathers wrote in April about an early warning system in Wisconsin designed to identify students at risk of dropping out? Well, the point of systems like these is to help schools figure out where to focus limited resources. They use past data about individual students, their schools, and districts to make predictions about current students. In Wisconsin, the system tends to be used by high schools considering how to serve incoming freshmen. But Todd cited some research out of the University of California, Berkeley in his story that looked at 10 years’ worth of data from Wisconsin’s system and found being labeled “high risk”—and therefore targeted for additional services—may not have made a difference.

Kristin Woelfel, policy counsel at the Center for Democracy & Technology and lead author of this week’s report, said those findings prompted alarm bells. Government entities have to be careful about where they take race into account when treating people differently. Woelfel told me there’s a basic standard in civil rights law: Use of race must be “narrowly tailored to achieve a compelling government interest.” 

The Wisconsin system ostensibly used race to help prevent dropouts. That sounds like a compelling government interest! But not if it didn’t work. The UC Berkeley researchers looked at students just above and below the cutoff for being labeled high risk, and expected to see that students labeled high risk got relatively more services and support, and therefore graduated at relatively higher rates—but they didn’t find that. “The “high risk” label may not have helped students at all: the researchers couldn’t rule out a 0 percent increase in graduation rates as a result of the label.

I discussed this with Cody Venzke, a senior policy counsel at the ACLU, which co-signed a letter the CDT sent to the U.S. Department of Education and the White House this week. The coalition urged the federal government to follow up its 2022 Blueprint for an AI Bill of Rights with more emphasis on civil rights. 

Venzke said it’s always worth being suspicious when race is used to make decisions about how to treat individuals, given historical discrimination. In Wisconsin, The Markup’s reporting made clear educators were not properly trained to understand the dropout risk labels and intervene accordingly. The investigation features students who said they found the labels stigmatizing. And if that stigma isn’t offset by demonstrable benefits for students—in higher graduation rates, for example—Venzke sees the potential for a civil rights violation.

But maybe it doesn’t have to be that way. Venzke points to schools’ obligation to protect students’ civil rights, ensure algorithmic systems are fair, and protect student privacy. 

“One of the ways you can achieve all three of those is by using aggregate data where aggregate data will fulfill the purpose,” Venzke said. 

One of the ways you can achieve all three of those is by using aggregate data where aggregate data will fulfill the purpose.

Cody Venzke, senior policy counsel at the ACLU

As it turns out, the UC Berkeley researchers found it would. In their paper, currently under peer review, they recommend Wisconsin scrap its use of individual factors like a student’s race or even test scores in determining that student’s dropout risk. Juan Perdomo, lead author on the paper and now a postdoctoral fellow at Harvard, told me that in highly segregated Wisconsin, building an early warning system with data from entire schools or districts offers equally reliable predictions about which students are likely to drop out. 

“It also provides a much more actionable way of improving graduation outcomes,” Perdomo told me. 

And isn’t that the point? Early warning systems aren’t supposed to just be statistical exercises in identifying students at risk. They’re supposed to help better serve those students. If an early warning system identifies entire schools as high risk—because it doesn’t factor in individual characteristics—maybe it would drive statewide spending or districtwide training toward those schools and create a clearer path toward making a difference.

For its part, the Wisconsin Department of Public Instruction is studying its early warning systems and considering changes. Communications Officer Chris Bucher told me the department has heard about concerns for some time now. So, we’ll see. 

In the meantime, we at The Markup will continue to report on algorithms that fuel discrimination and what that means for people’s lives.  

Thanks for reading! 

Tara García Mathewson
Reporter
The Markup

We don't only investigate technology. We instigate change.

Your donations power our award-winning reporting and our tools. Together we can do more. Give now.

Donate Now