Miami Herald (Sunday)

Worried about AI? How California lawmakers plan to tackle the technology’s risks in 2024

- BY QUEENIE WONG

Jodi Long was caught off guard by the cage filled with cameras meant to capture images of her face and body.

“I was a little freaked out because, before I walked in there, I said I don’t remember this being in my contract,” the actor said.

The filmmakers needed her digital scan, Long was told, because they wanted to make sure her arms were positioned correctly in a scene where she holds a computer-generated character.

That moment in 2020 stuck with Long, president of SAG-AFTRA’s Los Angeles local, while she was negotiatin­g for protection­s around the use of artificial intelligen­ce when actors went on strike. In November, the actors guild reached a deal with Hollywood studios that – among other things – required consent and compensati­on for the use of a worker’s digital replica.

Labor unions aren’t the only ones trying to limit AI’s potential threats.

Along with Gov. Gavin Newsom signing an executive order on AI in September, California lawmakers have introduced a raft of legislatio­n that sets the stage for more regulation in 2024. Some of the proposals focus on protecting workers, combating AI systems that can contribute to gender and racial biases and establishi­ng new requiremen­ts to safeguard against the misuse of AI for cybercrime­s, weapon developmen­t and propaganda.

Whether California lawmakers will succeed in passing AI legislatio­n, though, remains unclear. They’ll face lobbying from multibilli­on-dollar tech companies including Microsoft, Google and Facebook, political powerhouse­s that successful­ly stalled several AI bills introduced in 2023.

Artificial intelligen­ce has been around for decades. But as technology rapidly advances, the ability of machines to perform tasks associated with human intelligen­ce has raised questions about whether AI will replace jobs, fuel the spread of misinforma­tion or even lead to humanity’s extinction.

As lawmakers attempt to regulate AI, they’re also trying to understand how the technology works so they don’t hinder its potential benefits while simultaneo­usly trying to mitigate dangers.

“One of the core challenges is that this technology is dual use, meaning the same kind of technology that can, for instance, lead to massive improvemen­ts in healthcare can also be used potentiall­y to do pretty serious harm,” said Daniel Ho, a professor at Stanford University’s law school who advises the White House on AI policy.

Politician­s are feeling a sense of urgency, pointing to the resistance they’ve faced already in trying to control some of the mental health and child safety issues exacerbate­d by social media and other tech products. While some tech executives say they don’t oppose regulation, they’ve also said critics are exaggerati­ng the risks and expressed concern that they’ll have to deal with a patchwork of rules that vary around the world.

TechNet – a trade group that includes a variety of companies such as Apple, Google and Amazon – outlines on its website what members would and wouldn’t support when it comes to AI regulation. For example, TechNet says

policymake­rs should avoid “blanket prohibitio­ns on artificial intelligen­ce, machine learning, or other forms of automated decision-making” and not force AI developers to share informatio­n publicly that is proprietar­y.

State Assemblyme­mber Ash Kalra (D-San Jose) said policymake­rs don’t trust tech companies to regulate themselves.

“As a lawmaker, my intention is to protect the public and protect workers and protect against risks that may be created through unregulate­d AI,” Kalra said. “Those that are in the industry have different priorities.”

AI could affect 300 million full-time jobs, according to an April report by Goldman Sachs.

In September, Kalra introduced legislatio­n that would give actors, voice artists and other workers a way to nullify vague contracts that allow studios and other companies to use artificial intelligen­ce to digitally clone their voices, faces and bodies. Kalra said

he has no plans for now to set aside the bill, which is backed by SAG-AFTRA.

Federal lawmakers also have introduced legislatio­n aimed at protecting the voices and likenesses of workers. President Joe Biden signed an executive order on AI in October, noting how the technology could improve productivi­ty but also displace workers.

Duncan Crabtree-Ireland, the national executive director and chief negotiator of SAG-AFTRA, said he thinks it’s important that both state and federal lawmakers regulate AI without delay.

“It has to come from a variety of sources and [be] put together in a way that creates the ultimate picture that we all want to see,” he said.

Policymake­rs outside of the U.S. already have been moving forward. In December, the European Parliament and EU member states reached a landmark deal on the AI Act, calling the proposal “the world’s first comprehens­ive AI law.” The legislatio­n includes

a different set of rules based on how risky AI systems are and would also require AI tools that generate text, images and other content like OpenAI’s ChatGPT to publish what copyrighte­d data were used to train the systems.

As federal and state lawmakers fine-tune legislatio­n, workers are seeing how AI is affecting their jobs and testing whether current laws offer enough protection­s.

Tech companies – including Microsoft-backed OpenAI, Stability AI, Facebook parent Meta and Anthropic – are facing lawsuits over allegation­s that they used copyrighte­d work from artists and writers to train their AI systems. On Wednesday, the New York Times filed a lawsuit against Microsoft and OpenAI accusing the tech companies of using copyrighte­d work to create AI products that would compete with the news outlet.

Tim Friedlande­r, president and co-founder of the National Associatio­n of Voice Actors, said his members are losing out on jobs because some companies have decided to use AIgenerate­d voice. Actors have also alleged their voices are being cloned without their consent or compensati­on, a problem musicians face as well.

“One of the difficult things right now is that there’s no way to prove that something is human or synthetic or to be able to prove where the voice came from,” he said.

Worker protection­s are just one issue surroundin­g AI that California lawmakers will try to tackle in 2024.

Sen. Scott Wiener (DSan Francisco) in September introduced the Safety in Artificial Intelligen­ce Act, which aims to address some of the biggest risks posed by AI, he said, including the technology’s potential misuse in chemical and nuclear weapons, election interferen­ce and cyberattac­ks. Even though lawmakers don’t want to “squelch innovation,” they also want to be proactive, Wiener said.

“If you don’t get ahead of it, then it can be too late and we’ve seen that with social media and other areas where we should have been setting up at least broad stroke regulatory systems before the problem starts,” he said.

Lawmakers are also worried that AI systems could make mistakes that lead to unequal treatment of people based on protected characteri­stics such as race and gender. Assemblyme­mber Rebecca Bauer-Kahan (D-Orinda) is sponsoring a bill that would bar a person or entity from deploying an AI system or service that’s involved in making “consequent­ial decisions” that result in “algorithmi­c discrimina­tion.”

Concern that algorithms can amplify gender and racial biases because of what data are used to train the computer systems has been an ongoing issue in the tech industry.

 ?? ANDREW CABALLERO-REYNOLDS/AFP TNS ?? U.S. President Joe Biden and California Gov. Gavin Newsom take part in an event discussing the opportunit­ies and risks of Artificial Intelligen­ce at the Fairmont Hotel in San Francisco on June 20.
ANDREW CABALLERO-REYNOLDS/AFP TNS U.S. President Joe Biden and California Gov. Gavin Newsom take part in an event discussing the opportunit­ies and risks of Artificial Intelligen­ce at the Fairmont Hotel in San Francisco on June 20.

Newspapers in English

Newspapers from United States