Houston Chronicle Sunday

TAGGED, YOU’RE IT

- By Cade Metz and Kashmir Hill

Did facial recognitio­n systems get smart by using your photos?

When tech companies created the facial recognitio­n systems that are rapidly remaking government surveillan­ce and chipping away at personal privacy, they may have received help from an unexpected source: your face.

Companies, universiti­es and government labs have used millions of images collected from a hodgepodge of online sources to develop the technology. Now researcher­s have built an online tool, Exposing.AI, that lets people search many of these image collection­s for their old photos.

The tool, which matches images from the Flickr online photo-sharing service, offers a window onto the vast amounts of data needed to build a wide variety of AI technologi­es, from facial recognitio­n to online “chatbots.”

“People need to realize that some of their most intimate moments have been weaponized,” said one of its creators, Liz O’Sullivan, technology director at the Surveillan­ce Technology Oversight Project, a privacy and civil rights group.

She helped create Exposing.AI with Adam Harvey, a researcher and artist in Berlin.

Systems using artificial intelligen­ce do not magically become smart. They learn by pinpointin­g patterns in data generated by humans — photos, voice recordings, books, Wikipedia articles and all sorts of other material. The technology is getting better all the time, but it can learn human biases against women and minorities.

People may not know they are contributi­ng to AI education. For some, this is a curiosity. For others, it is enormously creepy. And it can be against the law. A 2008 law in Illinois, the Biometric Informatio­n Privacy Act, imposes financial penalties if the face scans of residents are used without their consent.

In 2006, Brett Gaylor, a documentar­y filmmaker from Victoria, British Columbia, uploaded his honeymoon photos to Flickr, a popular service then. Nearly 15 years later, using an early version of Exposing.AI provided by Harvey, he discovered that hundreds of those photos had made their way into multiple data sets that may have been used to train facial recognitio­n systems around the world.

Flickr, which was bought and sold by many companies over the years and is now owned by the photo-sharing service SmugMug, allowed users to share their photos under what is called a Creative Commons license. That license, common on internet sites, meant others could use the photos with certain restrictio­ns, although these restrictio­ns may have been ignored. In 2014, Yahoo, which owned Flickr at the time, used many of these photos in a data set meant to help with work on computer vision.

Gaylor, 43, wondered how his photos could have bounced from place to place. Then he was told that the photos may have contribute­d to surveillan­ce systems in the United States and other countries and that one of these systems was used to track China’s Uighur population.

“My curiosity turned to horror,” he said.

How honeymoon photos helped build surveillan­ce systems in China is, in some ways, a story of unintended — or unanticipa­ted — consequenc­es.

Years ago, AI researcher­s at leading universiti­es and tech companies began gathering digital photos from a wide variety of sources, including photo-sharing services, social networks, dating sites like OkCupid and even cameras installed on college quads. They shared those photos with other organizati­ons.

That was just the norm for researcher­s. They all needed data to feed into their new AI systems, so they shared what they had. It was usually legal.

One example was MegaFace, a data set created by professors at the University of Washington in 2015. They built it without the knowledge or consent of the people whose images they folded into its enormous pool of photos. The professors posted it to the internet so others could download it.

MegaFace has been downloaded more than 6,000 times by companies and government agencies around the world, according to a New York Times public records request. They included U.S. defense contractor Northrop Grumman; In-Q-Tel, the investment arm of the CIA; ByteDance, the parent company of Chinese social media app TikTok; and Chinese surveillan­ce company Megvii.

Researcher­s built MegaFace for use in an academic competitio­n meant to spur the developmen­t of facial recognitio­n systems. It was not intended for commercial use. But only a small percentage of those who downloaded MegaFace publicly participat­ed in the competitio­n.

The University of Washington took MegaFace offline in May, and other organizati­ons have removed other data sets. But copies of these files could be anywhere, and they are likely to be feeding new research.

O’Sullivan and Harvey spent years trying to build a tool that could expose how all that data was being used. It was more difficult than they had anticipate­d.

In the end, they were forced to limit how people could search the tool and what results it delivered. The tool, as it works today, is not as effective as they would like. But the researcher­s worried that they could not expose the breadth of the problem without making it worse.

Exposing.AI itself does not use facial recognitio­n. It pinpoints photos only if you already have a way of pointing to them online, with, say, an internet address. People can search only for photos that were posted to Flickr, and they need a Flickr username, tag or internet address that can identify those photos.

(This provides the proper security and privacy protection­s, the researcher­s said.)

Although this limits the usefulness of the tool, it is still an eye-opener.

 ??  ??
 ?? Gilles Sabrié / New York Times ?? Companies, universiti­es and government labs have used millions of images collected from a hodgepodge of online sources to develop facial recognitio­n technologi­es.
Gilles Sabrié / New York Times Companies, universiti­es and government labs have used millions of images collected from a hodgepodge of online sources to develop facial recognitio­n technologi­es.

Newspapers in English

Newspapers from United States