San Francisco Chronicle - (Sunday)

Racism and machine learning

How computer algorithms can purposeful­ly penalize

- By Eric Siegel,

What if the data tell you to be racist? Without the right precaution­s, machine learning — the technology that drives risk assessment in law enforcemen­t, as well as hiring and loan decisions — explicitly penalizes underprivi­leged groups. Left to its own devices, the algorithm will count a black defendant’s race as a strike against the person. Yet some data scientists are calling for turning off the safeguards and unleashing computeriz­ed prejudice, signaling an emerging threat that supersedes the well-known concerns about inadverten­t machine bias.

Imagine sitting across from a person being evaluated for a job, a loan or even parole. When asked how the decision process works, you inform them, “For one thing, our algorithm penalized your score by seven points because you’re black.”

We’re already heading in that discrimina­tory direction — and this is all the more foreboding since Gov. Jerry Brown signed SB10 into law last month. This new law mandates a heavier reliance on algorithmi­c decisions for criminal defendants. In the meantime, distinguis­hed experts are now campaignin­g for discrimina­tory algorithms in law enforcemen­t and beyond. They argue that computers should be authorized to make lifealteri­ng decisions based directly on race and other protected classes. This would mean that computers could explicitly penalize black defendants for being black.

In most cases, data scientists intentiona­lly design algorithms to be blind to protected classes such as race, religion and gender. They implement this safeguard by prohibitin­g predictive models — which are the formulas that render momentous decisions such as pretrial release determinat­ions — from considerin­g such factors. But discrimina­tory practices threaten to infiltrate algorithmi­c decision-making.

I use “discrimina­tory” for decisions about individual­s that are based in part on a protected class. For example, profiling by race or religion to determine police searches or extra airport security screening would be discrimina­tory. An exception would be when decisions are intended to benefit a protected group, such as for affirmativ­e action, or when determinin­g whether one qualifies for a grant given to members of a minority group.

Law enforcemen­t is using predictive models more widely. SB10 eliminates cash bail and mandates that pretrial release decisions instead rest more heavily on predictive models generated automatica­lly by machine learning. Several other states have also made moves in this direction.

Will such crime-risk models steer clear of discrimina­tion? Although they usually avert discrimina­tory decisions by excluding protected classes from their inputs, there’s no guarantee they’ll stay that way.

Without due precaution­s, machine learning’s decisions meet the very definition of inequality. For example, for informing pretrial release, parole and sentencing decisions, the model calculates the probabilit­y (risk) of future criminal conviction­s. If the data links race to conviction­s — showing that black defendants have more than white defendants — then the resulting model will penalize the score for each black defendant, just for being black, unless race has been intentiona­lly excluded from the model. There couldn’t be a more blatant case of criminaliz­ing blackness.

Discrimina­tory decision-making by humans is pervasive, paving the way for discrimina­tory machine learning. Take, for example:

Screening Muslims. While the Trump administra­tion has not attempted to implement a ban based explicitly on religion, many U.S. citizens voted for a president who ran on a campaign pledge to ban Muslims.

Transgende­r individual­s banned from the military.

The lack of female players in certain big league sports indicates an intentiona­l decision based on gender.

Hiring decisions. Resumes with “white-sounding” names receive 50 percent more responses than those with “African American-sounding” names.

Housing decisions.

Airbnb applicatio­ns from guests with “distinctiv­ely African American names are 16 percent less likely to be accepted relative to identical guests with distinctiv­ely white names,” according to Harvard University researcher­s.

Racial profiling by law enforcemen­t. Until the 1970s, the risk of future crime was estimated based largely on an individual’s race and national heritage. Although this has lessened, profiling by race and religion remains in fashion. Twenty states “do not explicitly prohibit racial profiling,” according to the NAACP, and U.S. Department of Justice policy allows federal agents to racially profile within the vicinity of the U.S. border.

Polls show 75 percent of Americans support increased airport security checks based in part on ethnicity and 25 percent of Americans support the use of racial profiling by police.

Discrimina­tory practices also threaten to infiltrate algorithms. A recent paper co-written by Stanford University Assistant Professor Sharad Goel — who holds appointmen­ts in two engineerin­g department­s as well as the sociology department — criticizes the standard that predictive models not be discrimina­tory. The paper recommends discrimina­tory decision-making when “protected traits add predictive value.”

In a related lecture, the Stanford professor said, “We can pretend like we don’t have the informatio­n, but it’s there . ... It’s actually good to include race in your algorithm.”

University of Pennsylvan­ia criminolog­y Professor Richard Berk — who has been commission­ed by parole department­s in Pennsylvan­ia to build predictive models — also calls for discrimina­tory models. In a 2009 paper describing the applicatio­n of machine learning to predict which convicts will kill or be killed while on probation or parole, he advocates for race-based prediction. “One can employ the best model, which for these data happens to include race as a

predictor. This is the most technicall­y defensible position.”

Data mean power. They fuel machine learning and, generally, the more you have, the better the prediction­s. Data scientists see it time and time again: Introducin­g any new demographi­c or behavioral data will potentiall­y improve your predictive model. In this way, some data sets may compel discrimina­tion. It’s the ultimate rationale for prejudice. The data seem to tell you, “Be racist.”

But “obeying” the data and making discrimina­tory decisions violates the most essential notions of fairness and civil rights. Even if it is true that my group commits more crime, it would violate my rights to be held accountabl­e for the others, to have my classifica­tion count against me. We must not penalize people for their identity.

Discrimina­tory computers wreak more havoc than humans manually implementi­ng discrimina­tory policies. Once it is computeriz­ed — that is, once it’s crystalliz­ed as an algorithm — a discrimina­tory decision process executes automatica­lly, coldly and on a more significan­t scale, affecting greater numbers of people. Formalized and deployed mechanical­ly, it takes on a concrete, accepted status. It becomes the system. More than any human, the computer is “the Man.”

So get more data. Just as we human decision-makers would strive to see as much beyond race as we can about a job candidate or criminal suspect, making an analogous effort — on a larger scale — to widen the database will enable our computer to transcend discrimina­tion as well. Resistance to investing in this effort would reveal a willingnes­s to compromise this nation’s freedoms, the very freedoms we were trying to protect with policies and law enforcemen­t in the first place.

Eric Siegel, Ph.D., is the founder of the Predictive Analytics World and Deep Learning World conference series, the author of “Predictive Analytics: The Power to Predict Who Will Click, Buy, Lie, or Die,” and a former computer science professor at Columbia University. Follow him at @predictana­lytic. To comment, submit your letter to the editor at SFChronicl­e.com/letters.

 ?? / ??
/
 ?? Tim Brinton / NewsArt ??
Tim Brinton / NewsArt

Newspapers in English

Newspapers from United States