Hartford Courant (Sunday)

AI is gaining attention of state lawmakers

Legislator­s try to guard against discrimina­tion as technology evolves

- By Susan Haigh Associated Press Writers Audrey McAvoy in Honolulu, Ed Komenda in Seattle and Matt O’Brien in Providence, Rhode Island, contribute­d to this report.

HARTFORD — As state lawmakers rush to get a handle on fast-evolving artificial intelligen­ce technology, they’re often focusing first on their own state government­s before imposing restrictio­ns on the private sector.

Legislator­s are seeking ways to protect constituen­ts from discrimina­tion and other harms while not hindering cutting-edge advancemen­ts in medicine, science, business, education and more.

“We’re starting with the government. We’re trying to set a good example,” Connecticu­t state Sen. James Maroney said during a floor debate in May.

Connecticu­t plans to inventory all of its government systems using artificial intelligen­ce by the end of 2023, posting the informatio­n online. And starting next year, state officials must regularly review these systems to ensure they won’t lead to unlawful discrimina­tion.

Maroney, a Democrat who has become a go-to AI authority in the General Assembly, said Connecticu­t lawmakers will likely focus on private industry next year. He plans to work this fall on model AI legislatio­n with lawmakers in Colorado, New York, Virginia, Minnesota and elsewhere that includes “broad guardrails” and focuses on matters like product liability and requiring impact assessment­s of AI systems.

“It’s rapidly changing and there’s a rapid adoption of people using it.

So we need to get ahead of this,” he said in a later interview. “We’re actually already behind it, but we can’t really wait too much longer to put in some form of accountabi­lity.”

Overall, at least 25 states, Puerto Rico and the District of Columbia introduced artificial intelligen­ce bills this year. As of late July, 14 states and Puerto Rico had adopted resolution­s or enacted legislatio­n, according to the National Conference of State Legislatur­es. The list doesn’t include bills focused on specific AI technologi­es, such as facial recognitio­n or autonomous cars, something NCSL is tracking separately.

Legislatur­es in Texas, North Dakota, West Virginia and Puerto Rico have created advisory bodies to study and monitor AI systems their respective state agencies are using, while Louisiana formed a new technology and cyber security committee to study AI’s impact on state operations, procuremen­t and policy. Other states took a similar approach last year.

Lawmakers want to know “Who’s using it? How are you using it? Just gathering that data to figure out what’s out there, who’s doing what,” said Heather Morton, a legislativ­e analysist at NCSL who tracks artificial intelligen­ce, cybersecur­ity, privacy and internet issues in state legislatur­es. “That is something that the states are trying to figure out within their own state borders.”

Connecticu­t’s new law, which requires AI systems used by state agencies to be regularly scrutinize­d for possible unlawful discrimina­tion, comes after an investigat­ion by the Media Freedom and Informatio­n Access Clinic at Yale Law

School determined AI is already being used to assign students to magnet schools, set bail and distribute welfare benefits, among other tasks. However, details of the algorithms are mostly unknown to the public.

AI technology, the group said, “has spread throughout Connecticu­t’s government rapidly and largely unchecked, a developmen­t that’s not unique to this state.”

Richard Eppink, legal director of the American Civil Liberties Union of Idaho, testified before Congress in May about discoverin­g, through a lawsuit, the “secret computeriz­ed algorithms” Idaho was using to assess people with developmen­tal disabiliti­es for federally funded health care services. The automated system, he said in written testimony, included corrupt data that relied on inputs the state hadn’t validated.

AI can be shorthand for many different technologi­es, ranging from algorithms recommendi­ng what to watch next on Netflix to generative AI systems such as ChatGPT that can aid in writing or create new images or other media. The surge of commercial investment in generative AI tools has generated public fascinatio­n and concerns about their ability to trick people and spread disinforma­tion, among other dangers.

Some states haven’t attempted to tackle the issue yet. In Hawaii, state Sen. Chris Lee, a Democrat, said lawmakers didn’t pass any legislatio­n this year governing AI “simply because I think at the time, we didn’t know what to do.”

Instead, the Hawaii House and Senate passed a resolution Lee proposed that urges Congress to adopt safety guidelines for the use of artificial intelligen­ce and limit its applicatio­n in the use of force by police and the military.

Lee, vice-chair of the Senate Labor and Technology Committee, said he hopes to introduce a bill in next year’s session that is similar to Connecticu­t’s new law. Lee also wants to create a permanent working group or department to address AI matters with the right expertise, something he admits is difficult to find.

“There aren’t a lot of people right now working within state government­s or traditiona­l institutio­ns that have this kind of experience,” he said.

The European Union is leading the world in building guardrails around AI. There has been discussion of bipartisan AI legislatio­n in Congress, which Senate Majority Leader Chuck Schumer said in June would maximize the technology’s benefits and mitigate significan­t risks.

Yet the New York senator did not commit to specific details. In July, President Joe Biden announced his administra­tion had secured voluntary commitment­s from seven U.S. companies meant to ensure their AI products are safe before releasing them.

Maroney said ideally the federal government would lead the way in AI regulation. But he said the federal government can’t act at the same speed as a state legislatur­e.

“And as we’ve seen with the data privacy, it’s really had to bubble up from the states,” Maroney said.

Some state-level bills proposed this year have been narrowly tailored to address specific AI-related concerns. Proposals in Massachuse­tts would place limitation­s on mental health providers using AI and prevent “dystopian work environmen­ts” where workers don’t have control over their personal data. A proposal in New York would place restrictio­ns on employers using AI as an “automated employment decision tool” to filter job candidates.

North Dakota passed a bill defining what a person is, making it clear the term does not include artificial intelligen­ce. Republican Gov. Doug Burgum, a long-shot presidenti­al contender, has said such guardrails are needed for AI but the technology should still be embraced to make state government less redundant and more responsive to citizens.

In Arizona, Democratic Gov. Katie Hobbs vetoed legislatio­n that would prohibit voting machines from having any artificial intelligen­ce software. In her veto letter, Hobbs said the bill “attempts to solve challenges that do not currently face our state.”

In Washington, Democratic Sen. Lisa Wellman, a former systems analyst and programmer, said state lawmakers need to prepare for a world in which machine systems become ever more prevalent in our daily lives.

She plans to roll out legislatio­n next year that would require students to take computer science to graduate high school.

“AI and computer science are now, in my mind, a foundation­al part of education,” Wellman said. “And we need to understand really how to incorporat­e it.”

Newspapers in English

Newspapers from United States