New York Daily News

Lack of city rules for AI risks ‘biased outcomes’

- BY MICHAEL GARTLAND

Several city government agencies risk “biased outcomes” due to a lack of guidelines for their use of artificial intelligen­ce, according to a new state comptrolle­r’s report.

The audit, issued Thursday morning by state Comptrolle­r Tom DiNapoli, focused primarily on four city agencies: the NYPD, the Administra­tion for Children’s Services and the Education and Buildings department­s.

It concluded that the city isn’t assessing the effectiven­ess of its AI tools or setting discernabl­e accuracy standards, “leaving it vulnerable to misguided, inaccurate or biased outcomes in several programs that can directly impact New Yorkers’ lives,” the comptrolle­r’s report said.

“Government’s use of artificial intelligen­ce to improve public services is not new,” DiNapoli said. “But there needs to be formal guidelines governing its use, a clear inventory of what’s being used and why, and accuracy standards, so that the algorithms that can help train educators, identify potential criminal suspects, prioritize child abuse cases and inspect buildings don’t do more harm than good.”

The most recently available city data show that, as of 2021, city agencies reported using 21 “algorithmi­c tools,” which include those employing artificial intelligen­ce or machine learning techniques. Examples of such tools include the NYPD’s use of facial recognitio­n technology, which has come under fire for flaws when it comes to race and ethnicity, and ACS’s use of technology aimed at predicting which open child-abuse cases are most likely to lead to future and severe harm.

According to DiNapoli’s report, the city doesn’t have an “effective AI governance framework,” and that while agencies are required to report their use of artificial intelligen­ce, there’s little in the way of guardrails around how to actually use the technology.

Newspapers in English

Newspapers from United States