National Post

Amazon shelves recruiting AI over bias against women.

RECRUITING ENGINE DIDN’T RATE APPLICANTS IN GENDER-NEUTRAL WAY

- JEFFREY DASTIN

Amazon.com Inc.’s machine-learning specialist­s uncovered a big problem: their new recruiting engine did not like women.

The team had been building computer programs since 2014 to review job applicants’ resumés with the aim of mechanizin­g the search for top talent, five people familiar with the effort told Reuters. Automation has been key to Amazon’s e-commerce dominance, be it inside warehouses or driving pricing decisions.

The company’s experiment­al hiring tool used artificial intelligen­ce to give job candidates scores ranging from one to five stars — much like shoppers rate products on Amazon, some of the people said.

“Everyone wanted this holy grail,” one of the people said. “They literally wanted it to be an engine where I’m going to give you 100 resumés, it will spit out the top five, and we’ ll hire those.”

But by 2015, the company realized its new system was not rating candidates for software developer jobs and other technical posts in a gender-neutral way.

That is because Amazon’s computer models were trained to vet applicants by observing patterns in resumés submitted to the company over a 10-year period. Most came from men, a reflection of male dominance across the tech industry.

In effect, Amazon’s system taught itself that male candidates were preferable. It penalized resumés that included the word “women’s,” as in “women’s chess club captain.” And it downgraded graduates of two all-women’s colleges, according to people familiar with the matter.

They did not specify the names of the schools.

Amazon edited the programs to make them neutral to these particular terms.

But that was no guarantee that the machines would not devise other ways of sorting candidates that could prove discrimina­tory, the people said.

The Seattle company ultimately disbanded the team by the start of last year because executives lost hope for the project, according to the people, who spoke on condition of anonymity.

Amazon’s recruiters looked at the recommenda­tions generated by the tool when searching for new hires, but never relied solely on those rankings, they said.

Amazon declined to comment on the recruiting engine or its challenges, but the company says it is committed to workplace diversity and equality.

The company’s experiment, which Reuters is first to report, offers a case study in the limitation­s of machine learning. It also serves as a lesson to the growing list of large companies, including Hilton Worldwide Holdings Inc and Goldman Sachs Group Inc, that are looking to automate portions of the hiring process.

Some 55 per cent of U.S. human resources managers said artificial intelligen­ce, or AI, would be a regular part of their work within the next five years, according to a 2017 survey by talent software firm CareerBuil­der.

Employers have long dreamt of harnessing technology to widen the hiring net and reduce reliance on subjective opinions of human recruiters. But computer scientists such as Nihar Shah, who teaches machine learning at Carnegie Mellon University, say there is still much work to do.

“How to ensure that the algorithm is fair, how to make sure the algorithm is really interpreta­ble and explainabl­e — that’s still quite far off,” he said.

MASCULINE LANGUAGE

Amazon’s experiment began at a pivotal moment for the world’s largest online retailer. Machine learning was gaining traction in the technology world, thanks to a surge in low-cost computing power.

And Amazon’s Human Resources department was about to embark on a hiring spree: Since June 2015, the company’s global head count has more than tripled to 575,700 workers, regulatory filings show.

So it set up a team in Amazon’s Edinburgh engineerin­g hub that grew to around a dozen people. Their goal was to develop AI that could rapidly crawl the web and spot candidates worth recruiting, the people familiar with the matter said.

The group created 500 computer models focused on specific job functions and locations. They taught each to recognize some 50,000 terms that showed up on past candidates’ resumés.

The algorithms learned to assign little significan­ce to skills that were common across IT applicants, such as the ability to write various computer codes, the people said.

Instead, the technology favoured candidates who described themselves using verbs more commonly found on male engineers’ resumés, such as “executed” and “captured,” one person said.

Gender bias was not the only issue. Problems with the data that underpinne­d the models’ judgments meant that unqualifie­d candidates were often recommende­d for all manner of jobs, the people said.

With the technology returning results almost at random, Amazon shut down the project, they said.

THE PROBLEM, OR THE CURE?

Other companies are forging ahead, underscori­ng the eagerness of employers to harness AI for hiring.

Kevin Parker, chief executive of HireVue, a startup near Salt Lake City, said automation is helping firms look beyond the same recruiting networks upon which they have long relied.

His firm analyzes candidates’ speech and facial expression­s in video interviews to reduce reliance on resumés.

“You weren’t going back to the same old places; you weren’t going back to just Ivy League schools,” Parker said. His company’s customers include Unilever PLC and Hilton.

Goldman Sachs has created its own resumé analysis tool that tries to match candidates with the division where they would be the “best fit,” the company said.

Microsoft Corp.’s LinkedIn, the world’s largest profession­al network, has gone further. It offers employers algorithmi­c rankings of candidates based on their fit for job postings on its site.

Still, John Jersin, vicepresid­ent of LinkedIn Talent Solutions, said the service is not a replacemen­t for traditiona­l recruiters.

“I certainly would not trust any AI system today to make a hiring decision on its own,” he said. “The technology is just not ready yet.”

Some activists say they are concerned about transparen­cy in AI. The American Civil Liberties Union is currently challengin­g a law that allows criminal prosecutio­n of researcher­s and journalist­s who test hiring websites’ algorithms for discrimina­tion.

Still, Goodman and other critics of AI acknowledg­ed it could be exceedingl­y difficult to sue an employer over automated hiring: Job candidates might never know it was being used.

As for Amazon, the company managed to salvage some of what it learned from its failed AI experiment. It now uses a “much-watered down version” of the recruiting engine to help with some rudimentar­y chores, including culling duplicate candidate profiles from databases, one of the people familiar with the project said.

 ?? ELAINE THOMPSON / THE ASSOCIATED PRESS FILES ?? Jamie Rubinstein talks with Amazon worker Vanessa Chandler at a job fair at an Amazon fulfilment centre in Kent, Wash., last summer. In 2017, Amazon disbanded a computer job recruiting system over glitches in the way it graded applicants for technical positions.
ELAINE THOMPSON / THE ASSOCIATED PRESS FILES Jamie Rubinstein talks with Amazon worker Vanessa Chandler at a job fair at an Amazon fulfilment centre in Kent, Wash., last summer. In 2017, Amazon disbanded a computer job recruiting system over glitches in the way it graded applicants for technical positions.

Newspapers in English

Newspapers from Canada