The Hamilton Spectator

Is the Google algorithm sexist?

- JULIA CARPENTER

Fresh off the revelation that Google image searches for “CEO” only turn up pictures of white men, there’s new evidence that algorithmi­c bias is, alas, at it again.

In a paper published in April, a team of researcher­s from Carnegie Mellon University claim Google displays far fewer ads for high-paying executive j obs … if you’re a woman.

“I think our findings suggest that there are parts of the ad ecosystem where kinds of discrimina­tion are beginning to emerge and there is a lack of transparen­cy,” Carnegie Mellon professor Annupam Datta told Technology Review. “This is concerning from a societal standpoint.”

To come to those conclusion­s, Datta and his colleagues basically built a tool, called Ad Fisher, that tracks how user behaviour on Google influences the personaliz­ed Google ads that each user sees. Because that relationsh­ip is complicate­d and based on a lot of factors, the researcher­s used a series of fake accounts: theoretica­l job-seekers whose behaviour they could track closely.

That online behaviour — visiting job sites and nothing else — was the same for all the fake accounts. But some listed their sex as men and some as women.

The Ad Fisher team found that when Google presumed users to be male job seekers, they were much more likely to be shown ads for high-paying executive jobs. Google showed the ads 1,852 times to the male group — but just 318 times to the female group.

But how much of this is us and how much of this is baked into the algorithm? It’s a question that a lot of people are struggling to answer.

After all, algorithmi­c personaliz­ation systems, like the ones behind Google’s ad platform, don’t operate in a vacuum: They’re pro- grammed by humans and taught to learn from user behaviour. So the more we click or search or generally Internet in sexist, racist ways, the algorithms learn to generate those results and ads.

“It’s part of a cycle: How people perceive things affects the search results, which affect how people perceive things,” Cynthia Matuszek, a computer ethics professor at University of Maryland and co-author of a study on gender bias in Google image search results, told The Washington Post in April.

Google cautions that some other things could be going on here, too. The advertiser in question could have specified that the ad only be shown to certain users for a whole host of reasons.

“Advertiser­s can choose to target the audience they want to reach, and we have policies that guide the type of interest-based ads that are allowed,” reads a statement from Google.

The interestin­g thing about the f ake users in the Ad Fisher study, however, is they had entirely fresh search histories: In f act, the accounts used were more or less identical, except for their listed gender identity. That would seem to indicate either that advertiser­s are requesting high-paying job ads only display to men or that some type of bias has been programmed, if inadverten­tly, into Google’s ad-personaliz­ation system.

In either case, Datta, the Carnegie Mellon researcher, says there’s room for much more scholarshi­p, and scrutiny, here.

“Many important decisions in society these days are being made by algorithms,” he said. “These algorithms run inside of boxes that we don’t have access to the internal details of. The genesis of this project was that we wanted to peek inside this box a little to see if there are more undesirabl­e consequenc­es of this activity going on.”

 ??  ??

Newspapers in English

Newspapers from Canada