Houston Chronicle

Dealing with bias in artificial intelligen­ce

- By Craig S. Smith

Bias is an unavoidabl­e feature of life, the result of the necessaril­y limited view of the world that any single person or group can achieve. But social bias can be reflected and amplified by artificial intelligen­ce in dangerous ways, whether it be in deciding who gets a bank loan or who gets surveilled.

The New York Times spoke with three prominent women in AI to hear how they approach bias in this powerful technology. Daphne Koller is a co-founder of online education company Coursera, and founder and chief executive of Insitro, a company using machine learning to develop new drugs. Koller, an adjunct professor in the computer science department at Stanford University, spoke to bias through the lens of machine-learning models.

Olga Russakovsk­y is an assistant professor in the department of computer science at Princeton University who specialize­s in computer vision and a co-founder of the AI4ALL foundation that works to increase diversity and inclusion within AI. Russakovsk­y is working to reduce bias in ImageNet, the data set that started the current machine-learning boom.

Timnit Gebru is a research scientist at Google on the ethical AI team and a co-founder of Black in AI, which promotes people of color in the field. Gebru has been instrument­al in moving a major internatio­nal AI conference, the Internatio­nal Conference on Learning Representa­tions, to Ethiopia next year after more than half of the Black in AI speakers could not get visas to Canada for a conference in 2018. She talked about the foundation­al origins of bias and the larger challenge of changing the scientific culture.

Their comments have been edited and condensed. Daphne Koller

You could mean bias in the sense of racial bias, gender bias. For example, you do a search for CEO on Google Images, and up come 50 images of white males and one image of CEO Barbie. That’s one aspect of bias.

Another notion of bias, one that is highly relevant to my work, are cases in which an algorithm is latching onto something that is meaningles­s and could potentiall­y give you very poor results. For example, imagine that you’re trying to predict fractures from X-ray images in data from multiple hospitals. If you’re not careful, the algorithm will learn to recognize which hospital generated the image. Some X-ray machines have different characteri­stics in the image they produce than other machines, and some hospitals have a much larger percentage of fractures than others. And so, you could actually learn to predict fractures pretty well on the data set that you were given simply by recognizin­g which hospital did the scan, without actually ever looking at the bone. The algorithm is doing something that appears to be good but is actually doing it for the wrong reasons. The causes are the same in the sense that these are all about how the algorithm latches onto things that it shouldn’t latch onto in making its prediction.

To recognize and address these situations, you have to make sure that you test the algorithm in a regime that is similar to how it will be used in the real world. So, if your machine-learning algorithm is one that is trained on the data from a given set of hospitals, and you will only use it in those same set of hospitals, then latching onto which hospital did the scan could well be a reasonable approach. It’s effectivel­y letting the algorithm incorporat­e prior knowledge about the patient population in different hospitals. The problem really arises if you’re going to use that algorithm in the context of another hospital that wasn’t in your data set to begin with. Then, you’re asking the algorithm to use these biases that it learned on the hospitals that it trained on, on a hospital where the biases might be completely wrong.

Overall, there’s not nearly as much sophistica­tion as there needs to be out there for the level of rigor that we need in terms of the applicatio­n of data science to real-world data, and especially biomedical data. Olga Russakovsk­y

I believe there are three root causes of bias in artificial intelligen­ce systems. The first one is bias in the data. People are starting to research methods to spot and mitigate bias in data. For categories like race and gender, the solution is to sample better such that you get a better representa­tion in the data sets. But, you can have a balanced representa­tion and still send very different messages. For example, female programmer­s are frequently depicted sitting next to a man in front of the computer or with a man watching over her shoulder.

I think of bias very broadly. Certainly gender and race and age are the easiest to study, but there are all sorts of angles. Our world is not fair. There’s no balanced representa­tion of the world, and so data will always have a lot of some categories and relatively little of others.

Going further, the second root cause of bias is in the algorithms themselves. Algorithms can amplify the bias in the data, so you have to be thoughtful about how you actually build these systems.

This brings me to the third cause: human bias. AI researcher­s are primarily people who are male, who come from certain racial demographi­cs, who grew up in high socioecono­mic areas, primarily people without disabiliti­es. We’re a fairly homogeneou­s population, so it’s a challenge to think broadly about world issues. There are a lot of opportunit­ies to diversify this pool, and as diversity grows, the AI systems

themselves will become less biased.

Let me give one example illustrati­ng all three sources. The ImageNet data set was curated in 2009 for object recognitio­n, containing more than 14 million images. There are several things we are doing with an eye toward rebalancin­g this data set to better reflect the world at large. So far, we went through 2,200 categories to remove those that may be considered offensive. We’re working on designing an interface to let the community flag additional categories or images as offensive, allowing everyone to have a voice in this system. We are also working to understand the impact that such changes would have on the downstream computer vision models and algorithms.

I don’t think it’s possible to have an unbiased human, so I don’t see how we can build an unbiased AI system. But we can certainly do a lot better than we’re doing. Timnit Gebru A lot of times, people are talking about bias in the sense of equalizing performanc­e across groups. They’re not thinking about the underlying foundation, whether a task should exist in the first place, who creates it, who will deploy it on which population, who owns the data, and how is it used?

The root of these problems is not only technologi­cal. It’s social. Using technology with this underlying social foundation often advances the worst possible things that are happening. In order for technology not to do that, you have to work on the underlying foundation as well. You can’t just close your eyes and say: “Oh, whatever, the foundation, I’m a scientist. All I’m going to do is math.”

For me, the hardest thing to change is the cultural attitude of scientists. Scientists are some of the most dangerous people in the world because we have this illusion of objectivit­y; there is this illusion of meritocrac­y, and there is this illusion of searching for objective truth. Science has to be situated in trying to understand the social dynamics of the world because most of the radical change happens at the social level.

We need to change the way we educate people about science and technology. Science currently is taught as some objective view from nowhere (a term I learned about from reading feminist studies works), from no one’s point of view. But there needs to be a lot more interdisci­plinary work, and there needs to be a rethinking of how people are taught things.

People from marginaliz­ed groups have been working really hard to bring this to the forefront, and then once it’s brought to the forefront other people from nonmargina­lized groups start taking all the credit and pouring money into “initiative­s.” They’re not going to take the kinds of risks that people in marginaliz­ed communitie­s take because it’s not their community that’s being harmed.

All these institutio­ns are bringing the wrong people to talk about the social impacts of AI or be the faces of these things just because they’re famous and privileged and can bring in more money to benefit the already privileged.

There are some things that should be discussed on a global stage, and there should be agreements across countries. And there are other things that should just be discussed locally. We need to have principles and standards, and governing bodies, and people voting on things and algorithms being checked, something similar to the FDA [Food and Drug Administra­tion]. So, for me it’s not as simple as creating a more diverse data set and things are fixed. That’s just one component of the equation.

 ??  ??
 ?? Harriet Lee-Merrion / New York Times ?? ABOVE: Bias is an unavoidabl­e feature of life, the result of the limited view of the world that any single person or group can achieve. But social bias can be reflected and amplified by artificial intelligen­ce in dangerous ways.
ON THE COVER: Timnit Gebru, a research scientist at Google.
Harriet Lee-Merrion / New York Times ABOVE: Bias is an unavoidabl­e feature of life, the result of the limited view of the world that any single person or group can achieve. But social bias can be reflected and amplified by artificial intelligen­ce in dangerous ways. ON THE COVER: Timnit Gebru, a research scientist at Google.
 ?? David Crow / New York Times ?? Olga Russakovsk­y is an assistant professor in the Department of Computer Science at Princeton University.
David Crow / New York Times Olga Russakovsk­y is an assistant professor in the Department of Computer Science at Princeton University.
 ?? Daphne Koller / New York Times ?? Daphne Koller is an adjunct professor in the computer science department at Stanford University.
Daphne Koller / New York Times Daphne Koller is an adjunct professor in the computer science department at Stanford University.

Newspapers in English

Newspapers from United States