-
(单词翻译:双击或拖选)
ARI SHAPIRO, HOST:
In the fall of 2008, an Indiana woman named Omega Young got a letter saying she needed to recertify for the state's public benefits program.
VIRGINIA EUBANKS: But she was unable to make the appointment because she was suffering from ovarian cancer.
SHAPIRO: She called the local office to say she wouldn't make the appointment because she was hospitalized getting cancer treatments and she lost her benefits anyway. The reason - failure to cooperate.
EUBANKS: So because she lost her benefits, she couldn't afford her medications, she lost her food stamps, she couldn't pay her rent. She lost access to free transportation to her medical appointments. And Omega Young died on March 1, 2009. And on the next day, she won an appeal for wrongful termination and all of her benefits were restored the day after her death.
SHAPIRO: This is one of the stories the author Virginia Eubanks tells in her latest book "Automating1 Inequality: How High-Tech2 Tools Profile, Police, And Punish The Poor." That book is the subject of this week's All Tech Considered.
(SOUNDBITE OF MUSIC)
SHAPIRO: Virginia Eubanks argues that many of the automated3 systems that deliver public services today are rigged against the people these programs are supposed to serve. She dives deep into three examples of automated public services - welfare benefits in Indiana, housing for the homeless in Los Angeles and children's services in Allegheny County, Pa., which includes Pittsburgh.
The Indiana case was so bad that the state eventually gave up on the automated system. Virginia Eubanks started by telling me what state lawmakers were trying to accomplish through automation.
EUBANKS: Indiana was attempting to save money and to make the system more efficient. But the way the system rolled out, it seems like one of the intentions was actually to break the relationship between caseworkers and the families they served. The governor sort of did a press tour around this contract. And one of the things he kept bringing up was there was one case where two case workers had colluded with some recipients4 to defraud5 the government for about - I think it was about $8,000.
And the governor used this case over and over and over again to suggest that when caseworkers and families have personal relationships, that it's an invitation to fraud. So the system was actually designed to break that relationship. So what happened is the state replaced about 1,500 local caseworkers with online forms and regional call centers.
And that resulted in a million benefits denials in the first three years of the experiment, which was a 54 percent increase from the three years before.
SHAPIRO: Is an automated system of public services inherently going to be less helpful, less effective than something like Uber or Lyft or Amazon or all the automated things that people who are not in poverty rely on every day?
EUBANKS: No. There's nothing intrinsic in automation that makes it bad for the poor. One of my greatest fears in this work is that we're actually using these systems to avoid some of the most pressing moral and political challenges of our time, specifically poverty and racism6. So we're kind of using these systems as a kind of empathy override7. You know, let's talk about Los Angeles.
So there's 58,000 unhoused folks in Los Angeles. It's the second-highest population in the United States and 75 percent of them are completely unsheltered, which means they're just living in the street. I do not want to be the case worker who is making that decision, who is saying there's 50,000 people with no resources. I have, you know, a handful of resources available. Now I have to pick.
But the problem is that we are using these tools to basically outsource that incredibly hard decision to machines.
SHAPIRO: So the underlying8 problem is not that the housing system is automated but it sure doesn't help that automating that system allows people to ignore, more or less, the fact that there are not enough houses.
EUBANKS: Yeah. So one of the folks I talked to in the book, this great, brilliant man Gary Blasi has one of the best quotes in the book and he says, homelessness is not a systems engineering problem. It's a carpentry problem, right?
SHAPIRO: If you've got 10 houses for 20 people, it doesn't matter how good the system for housing those people is, it's not going to work.
EUBANKS: Exactly.
SHAPIRO: As you point out in the book, caseworkers have biases10. There are case workers who are racist11, who discriminate12, who favor some clients over others for inappropriate reasons. Doesn't automation have the potential to solve those problems?
EUBANKS: Yeah, let's be absolutely direct about this that human bias9 in public assistance systems have created deep inequalities for decades. And it's specifically around the treatment of black and brown folks, who have often been either overrepresented in the more punitive13 systems or diverted from the more helpful systems because of frontline caseworker bias.
SHAPIRO: So they get thrown in prison more often or their children taken away more often, they get public housing less often, that sort of thing.
EUBANKS: Exactly. But the thing that's really important to understand about the systems I profile in "Automating Inequality" is that these systems don't actually remove that bias, they simply move it. So in Allegheny County where I look at the predictive model that's supposed to be able to forecast which children will be victims of abuse or neglect in the future, in that case, one of the hidden biases is that it uses proxies14 instead of actual measures of maltreatment.
And one of the proxies it uses is called call re-referral, which just means that a child is called on and then a second call comes in within two years. And the problem with this is that both anonymous15 reporters and mandated16 reporters report black and biracial families for abuse and neglect 3.5 times more often than they report white families.
SHAPIRO: You draw these three detailed17 pictures of automated systems falling short in Indiana, California, Pennsylvania. Do you think a different author could have found three different automated systems somewhere in the country that were working really well in providing services effectively?
EUBANKS: Absolutely. One of the things that's different about the way that I wrote the book is that I started from the point of view of the targets of these systems. It doesn't mean I only spoke18 to those folks. But I spoke to, you know, unhoused folks, both those who have had luck getting housing through coordinated19 entry and those who haven't. I spoke to families who have been investigated for maltreatment.
And I will say that when you start from the point of view of these very vulnerable families, that these systems look really different than they look from the point of view of the data scientists or administrators20 who are developing them. And I wasn't hearing these voices at all in the debates that we've been having about what's sort of coming to be known as algorithmic accountability or algorithmic fairness.
I was never hearing the voices of the people who face the pointy end of the most punitive stick. And I really thought it was important to bring those stories to the table.
SHAPIRO: Virginia Eubanks, thanks so much for talking with us.
EUBANKS: Thank you so much.
SHAPIRO: Her book is called "Automating Inequality: How High-Tech Tools Profile, Police, And Punish The Poor."
1 automating | |
(使)自动化( automate的现在分词 ) | |
参考例句: |
|
|
2 high-tech | |
adj.高科技的 | |
参考例句: |
|
|
3 automated | |
a.自动化的 | |
参考例句: |
|
|
4 recipients | |
adj.接受的;受领的;容纳的;愿意接受的n.收件人;接受者;受领者;接受器 | |
参考例句: |
|
|
5 defraud | |
vt.欺骗,欺诈 | |
参考例句: |
|
|
6 racism | |
n.民族主义;种族歧视(意识) | |
参考例句: |
|
|
7 override | |
vt.不顾,不理睬,否决;压倒,优先于 | |
参考例句: |
|
|
8 underlying | |
adj.在下面的,含蓄的,潜在的 | |
参考例句: |
|
|
9 bias | |
n.偏见,偏心,偏袒;vt.使有偏见 | |
参考例句: |
|
|
10 biases | |
偏见( bias的名词复数 ); 偏爱; 特殊能力; 斜纹 | |
参考例句: |
|
|
11 racist | |
n.种族主义者,种族主义分子 | |
参考例句: |
|
|
12 discriminate | |
v.区别,辨别,区分;有区别地对待 | |
参考例句: |
|
|
13 punitive | |
adj.惩罚的,刑罚的 | |
参考例句: |
|
|
14 proxies | |
n.代表权( proxy的名词复数 );(测算用的)代替物;(对代理人的)委托书;(英国国教教区献给主教等的)巡游费 | |
参考例句: |
|
|
15 anonymous | |
adj.无名的;匿名的;无特色的 | |
参考例句: |
|
|
16 mandated | |
adj. 委托统治的 | |
参考例句: |
|
|
17 detailed | |
adj.详细的,详尽的,极注意细节的,完全的 | |
参考例句: |
|
|
18 spoke | |
n.(车轮的)辐条;轮辐;破坏某人的计划;阻挠某人的行动 v.讲,谈(speak的过去式);说;演说;从某种观点来说 | |
参考例句: |
|
|
19 coordinated | |
adj.协调的 | |
参考例句: |
|
|
20 administrators | |
n.管理者( administrator的名词复数 );有管理(或行政)才能的人;(由遗嘱检验法庭指定的)遗产管理人;奉派暂管主教教区的牧师 | |
参考例句: |
|
|