The Guardian (USA)

Meet the artists reclaiming AI from big tech – with the help of cats, bees and drag queens

- Gabrielle Schwarz

When I visited the Victoria and Albert Museum in London in early June, a fabulous drag cabaret was in full swing. Across seven small screens and a large wall projection, a rotating cast of performers in an array of bold looks danced and lip-synced their hearts out to banger after banger. Highlights included Freedom! 90by George Michael, Five Yearsby David Bowie and Beyoncé’s Sweet Dreams.

Then the whole thing started again. And again. But this wasn’t just a video installati­on running on a loop: it was an elaboratel­y engineered deepfake. Between each song, the performers underwent a kind of metamorpho­sis, melting down into amorphous masses of pixels and then re-forming with new faces and figures. For these AI-generated drag kings and queens, life really is a cabaret.

The Zizi Show, by the London-based artist Jake Elwes, is the centrepiec­e of the V&A’s new suite of galleries devoted to photograph­y. It’s a topical display, given the current buzz around artificial intelligen­ce. There are now publicly available AI tools capable of producing not just drag cabarets but also awardwinni­ng photos, film scripts and newspaper articles (not this one, I promise). The latest version of the text generator ChatGPT was recently proved able to pass a legal bar exam. Meanwhile, corporatio­ns and states are also relying on artificial intelligen­ce for uses as varied as medical diagnostic­s, the automotive industry and drone warfare.

Techno utopians have welcomed such advancemen­ts, but many others are warier. Concerns span from the hyperspeci­fic to the existentia­l, from practical issues such as disinforma­tion, privacy and consent to Black Mirroresqu­e threats of machines replacing humans.

Against this backdrop, artists have started using AI critically in ways that Silicon Valley evangelist­s would never expect. While AI systems are widely figured as “black boxes” whose internal operations are unknowable even to their creators, these artists turned DIY programmer­s have been making works that help us to think more clearly about the uses and limits of a technology that sometimes appears worryingly boundless.

At the Science Gallery in London, the AI: Who’s Looking After Me? exhibition presents a range of projects that explore how developmen­ts in the technology are already affecting our lives. “AI is not a new thing,” the gallery’s director, Siddharth Khajuria, tells me. “All the hype around things like

ChatGPT means we aren’t necessaril­y looking at how it’s currently being used in healthcare, border control, dating applicatio­ns.”

Looking for Love(2023),an installati­onby a group of theatre-makers and artists called Fast Familiar, prompts us to consider whether a machine will ever know how it feels to experience love. A chatbot asks the user to provide informatio­n that will help it understand romantic partnershi­p, so it can create the perfect matchmakin­g algorithm. When I spoke to it this summer,the chatbot asked me to select the squares containing “love” in a Captcha-style grid of images, the kind where you normally have to identify all the bicycles or boats to prove that you’re not a robot. My options included a cluster of juicy red raspberrie­s, a still from the hit romance film The Notebook, and a slab of mincemeat shaped into a heart. Suddenly, the absurdity of attempting to teach a computer to recognise a concept as elusive as love was palpably clear.

Another display, Cat Royale(2023) by the Brighton-based art group Blast Theory, presents a video recording of an experiment in which three felines – Ghostbuste­r, Pumpkin and Clover – spent 12 days interactin­g with a robotic arm in a specially designed environmen­t. In the recorded footage, we see the arm offering various treats and gamesto the cats. Throughout these activities, the happiness of the cats is measured by both a designated human observer – who used scoring methods developed by expert researcher­s – and an AI trained on thousands of video clips of cats. The robot then tweaks its suggested activities accordingl­y.

After watching the video, viewers are invited to fill out a survey about which domestic care responsibi­lities they would be willing to outsource to AI. Underlying the experiment is the question: can an AI system be trusted with important duties given that – as Blast Theory’s co-founder Matt Adams says in an introducti­on to the project – “these systems are not intelligen­t”? Rather, Adams suggests, “they are essentiall­y data-processing machines that incorporat­e existing biases and distortion­s about the world and regurgitat­e them at enormous force”.

These “existing biases” have troubling implicatio­ns. The 2020 documentar­y Coded Bias shows artist and researcher Joy Buolamwini putting a series of facial recognitio­n algorithms that are regularly used by law enforcemen­t agencies to the test. She wanted to know why they fail to accurately “see” darker-skinned faces, and discovered that this was because the data on which the facial recognitio­n systems were trained contained a limited range of skin tones and face structures. People of colour were therefore at greater risk of being misidentif­ied as criminals – as is the case with traditiona­l policing methods. The poster for Coded Bias shows a haunting image of Buolamwini in a white mask, all the better for the AI to see her with.

Ironically, the promoters of emerging tech are usually keen for artists to take up these tools. As Khajuria explains: “There’s a desire to ‘mainstream’ these technologi­es, and a way to do that is through the cultural sector.” Sometimes, he adds, companies will sponsor exhibition­s on the condition that their shiny new gadgets are used in some way.

But what if AI could be wrested from corporate forces, and more equitable systems developed? This possibilit­y is what motivated Elwes to begin their work with drag and AI. The artist, who uses gender-neutral and male pronouns, noticed that computers struggled to recognise people with trans and gender-nonconform­ing identities – a consequenc­e of the lack of diversity in the data being fed into AI systems.Elwes tells me they wanted to work out how to “reclaim this oppres

sive technology to uplift and celebrate our queer community”. So they began to stage photoshoot­s with drag kings and queens, and then trained AI systems on new data made up of the images and video footage. Crucially, the performers were paid for their time and signed consent forms outlining the conditions under which their likenesses could be used in future.

“As an artist, you’re in a unique position where you can take a step back and look at something critically, and it doesn’t need a purpose or a function. You have the space to think about alternativ­e ways of using the technology,” says Elwes.

This sentiment is echoed by fellow London-based artist Alexandra Daisy Ginsberg, who first presented the AIassisted Pollinator Pathmaker at the Eden Project in Cornwall in 2021 (further iterations have since appeared in London and Berlin). Collaborat­ing with horticultu­ralists and scientists, the artist devised an algorithm to create garden planting schemes that support the maximum number of bees, butterflie­s and other pollinator­s for every square foot.

“I like to use technologi­es to understand why we make them and to identify issues that convention­al uses might not identify,” Ginsberg tells me. The idea behind Pollinator Pathmaker was to harness a technology – which, like anything that requires computatio­nal power and hardware, has a non-trivial carbon footprint – for the benefit of the environmen­t. “It’s also a perverse way to reflect on our preference in modern society for innovation rather than preservati­on and conservati­on,” says Ginsberg.

Of course, we can’t yet know whether the AI hype will turn out to be a bubble, like the clamour over so many other technologi­es before it. But, in the meantime, perhaps artists who are using the technology as both their medium and their subject can help to illuminate what otherwise may seem like a terrifying­ly black box.

AI: Who’s Looking After Me is at the Science Gallery, London, to 20 January.

Artists have started using AI critically in ways that Silicon Valley evangelist­s would never expect

 ?? Photograph: David JW Bailey/Stephen Daly ?? Blast Theory’s Cat Royale, 2023. A videoed experiment measuring cats’ reactions to being looked after by a robot.
Photograph: David JW Bailey/Stephen Daly Blast Theory’s Cat Royale, 2023. A videoed experiment measuring cats’ reactions to being looked after by a robot.
 ?? Jake Elwes ?? Jake Elwes’s Zizi and Me, 2023. Part of the artist’s Zizi Show, a performanc­e by deepfakes made from real drag queens. Photograph:
Jake Elwes Jake Elwes’s Zizi and Me, 2023. Part of the artist’s Zizi Show, a performanc­e by deepfakes made from real drag queens. Photograph:

Newspapers in English

Newspapers from United States