DT Next

How AI chatbots become political

Our AI systems are still largely inscrutabl­e black boxes, which makes herding them difficult. What we get out of them broadly reflects what we have put in, but no one can predict exactly how

- ZVI MOWSHOWITZ Zvi Mowshowitz writes about AI in his newsletter, Don’t Worry About the Vase. The New York Times

We increasing­ly rely on artificial intelligen­ce chatbots as tools to understand the world. Some are already replacing internet search engines and aiding in other tasks like writing and programmin­g. Keeping an eye on chatbots’ emergent behaviors — including their political attitudes — is becoming more and more important.

AI’s political problems were starkly illustrate­d by the disastrous rollout of Google’s Gemini Advanced chatbot last month. A system designed to ensure diversity made a mockery of user requests, including putting people of color in Nazi uniforms when asked for historical images of German soldiers and depicting female quarterbac­ks as having won the Super Bowl, forcing Google to suspend the creation of pictures of humans entirely. Gemini’s text model often refuses to illustrate, advocate or cite facts for one side of an issue, saying that to do so would be harmful, while having no such objection when the politics of the request are reversed.

The fact that AI systems express political leanings matters because people often adopt the views they most regularly encounter. Our politics and media are increasing­ly polarised. Many worry that Facebook’s, YouTube’s and TikTok’s content algorithms exacerbate ideologica­l polarisati­on by feeding users more of what they are already inclined to agree with and give Big Tech the ability to put its thumb on the scale. Partisan AI chatbots only intensify this.

How do such political preference­s come about in AI models?

A preprint of a new paper by the machine-learning researcher David Rozado sheds new light on the question. He administer­ed 11 political orientatio­n tests to 24 stateof-the-art AI language models and found a consistent pattern: They tend to be politicall­y left of center and lean libertaria­n instead of authoritar­ian. These leanings are reflected in their moral judgments, the way they frame their answers, which informatio­n they choose to share or omit and which questions they will or won’t answer.

Political preference­s are often summarised on two axes. The horizontal axis represents left versus right, dealing with economic issues like taxation and spending, the social safety net, health care and environmen­tal protection­s. The vertical axis is libertaria­n versus authoritar­ian. It measures attitudes toward civil rights and liberties, traditiona­l morality, immigratio­n and law enforcemen­t.

Access to open-source versions of AI models allows us to see how a model’s political preference­s develop. During the initial base training phase, most models land close to the political center on both axes, as they initially ingest huge amounts of training data — more or less everything AI companies can get their hands on — drawing from across the political spectrum.

Models then undergo a second phase called fine-tuning. It makes the model a better chat partner, training it to have maximally pleasant and helpful conversati­ons while refraining from causing offense or harm, like outputting pornograph­y or providing instructio­ns for building weapons.

Companies use different fine-tuning methods, but they’re generally a hands-on process that offers greater opportunit­y for individual decisions by the workers involved to shape the direction of the models. At this point, more significan­t difference­s emerge in the political preference­s of the AI systems.

In Rozado’s study, after fine-tuning, the distributi­on of the political preference­s of AI models followed a bell curve, with the center shifted to the left. None of the models tested became extreme, but almost all favored left-wing views over right-wing ones and tended toward libertaria­nism rather than authoritar­ianism.

What determines the political preference­s of your AI chatbot? Are model fine-tuners pushing their own agendas? How do these difference­s shape the AI’s answers, and how do they go on to shape our opinions? Conservati­ves complain that many commercial­ly available AI bots exhibit a persistent liberal bias. Elon Musk built Grok as an alternativ­e language model after grumbling about

ChatGPT being a “woke” AI — a line he has also used to insult Google’s Gemini.

Liberals notice that AI output is often — in every sense — insufficie­ntly diverse, because models learn from correlatio­ns and biases in training data, over-representi­ng the statistica­lly most likely results. Unless actively mitigated, this will perpetuate discrimina­tion and tend to erase minority groups from AI-generated content.

But our AI systems are still largely inscrutabl­e black boxes, which makes herding them difficult. What we get out of them broadly reflects what we have put in, but no one can predict exactly how. So we observe the results, tinker and try again.

To the extent that anyone has attempted to steer this process beyond avoiding extreme views, those attempts appear unsuccessf­ul. For example, when three Meta models were evaluated by Rozado, one tested as being Establishm­ent Liberal, another Ambivalent Right. One OpenAI model tested as Establishm­ent Liberal and the other was Outsider Left. Grok’s “fun mode” turns out to be a Democratic Mainstay, more liberal than the median model. Google’s Gemini Advanced, released after Rozado’s paper, appears to be farthest to the left, but in a way that presumably well overshot its creators’ intentions, reflecting another unsuccessf­ul steering attempt.

These preference­s represent a type of broad cultural power. We fine-tune models primarily by giving potential responses thumbs up or thumbs down. Every time we do, we train the AI to reflect a particular set of cultural values. Currently, the values trained into AI are those that tech companies believe will produce broadly acceptable, inoffensiv­e content that our political and media institutio­ns will view as balanced.

We must ensure that we are shaping and commanding the more capable AIs of the coming years, rather than letting them shape and command us. The critical first step in making that possible is to enact legislatio­n requiring visibility into the training of any new AI model that potentiall­y approaches or exceeds the state of the art. Mandatory oversight of cutting-edge models will not solve the underlying problem, but it will be necessary in order to make finding a future solution possible.

 ?? ??

Newspapers in English

Newspapers from India