AugustMan (Malaysia)

Find out whether facial recognitio­n is it all it’s cracked up to be

Is facial recognitio­n all it’s cracked up to be? Our images can provide access to a repository of personal informatio­n – dare we risk that data being on any device other than our own?

- WORDS BY EVIGAN XIAO PHOTO BY GETTY IMAGES

IT WASN’T THAT LONG AGO when experts saw facial recognitio­n technology as the “it” thing in the biometric security landscape. The premise was simple and sensible enough: with the right algorithm, the unique compositio­n of any person’s face can be reduced to a digital key that’s used as a security token. It’s an incredible modern convenienc­e, but how do we make sure that these measures work as intended while remaining in the possession of their rightful owners?

The growing trend of utilising automated facial recognitio­n

(AFR) in a variety of sectors has raised numerous concerns. What started as a novel feature for unlocking smartphone­s started seeing applicatio­ns in more sensitive areas such as online banking, surveillan­ce and law enforcemen­t. Naturally, this has led to questions about the accuracy and legitimacy of AFR.

FACIAL NUANCES ARE TRICKY

For biometric security to be viable, it must be accurate. Unfortunat­ely, facial recognitio­n software is not immune to false positives. In fact, they can perform rather poorly: earlier in February, the British Metropolit­an Police deployed facial recognitio­n tech on 8,600 pedestrian­s in London (without consent, by the way, in a clear infringeme­nt of privacy). The system, which interfaced with the Met Police database, generated eight alerts. Here’s the kicker: only one of them was an accurate identifica­tion that led to an arrest. In other words, the error rate was a whopping 87.5 per cent.

AI experts have also highlighte­d the limitation­s of AFR systems when deployed across racial and gender lines. Dr Fanglin Wang, head of the artificial intelligen­ce unit at Advance.AI, stressed the importance of creating localised data sets when we spoke to him about this. “If facial recognitio­n technology is to work in Singapore or Southeast Asia, for instance, you need to train the underlying algorithm with the relevant local data,” he said. “That means exposing it to Singaporea­n or Southeast Asian faces, as their facial structures and skin tones are quite different from those of Caucasians.”

False positives and negatives are often the result of low quality images that are either blur or poorly lit. Dr Wang pointed to advancemen­ts in technology, such as better sensors and depth control on cameras, as being the solution to this problem. He also believes that AFR accuracy has the potential to supersede that of the human eye given “the right data”.

TRUSTING THE KEYMAKERS

And what about those who manage the software? Every security system comes with a backend. In the case of AFR technology, building and maintainin­g a database of images is crucial to operations. Therein lies the question: how will such a database be compiled? Policies on personal data such as Singapore’s Personal Data Protection Act impose strict guidelines on how personal informatio­n such as facial images can be gathered and used. As a result, AFR developers sometimes come under fire for using unsavoury methods to collect images.

On 10 March 2020, Vermont attorney general Thomas J Donovan filed a lawsuit against data broker Clearview AI over alleged privacy violations. The New York-based facial recognitio­n app company is currently under investigat­ion for creating a searchable database

containing billions of facial images harvested from social media platforms. The database also allowed users to upload specific facial images and view matches. Other individual­s in Illinois and New York have since joined in the lawsuit.

Clearview AI’s initial defence was that it only gave access to law enforcemen­t agencies. Investigat­ions, however, have revealed that users included corporatio­ns, wealthy individual­s and investors; American billionair­e John Catsimatid­is has openly admitted to using the Clearview AI app to spy on his daughter’s date in an article published by The New York Times.

In response to the scandal, sites such as YouTube, Facebook, Twitter and LinkedIn have issued cease-and-desist letters to Clearview AI demanding that it halt its data mining. Apple has also suspended the company from its developer programme.

TRADING PRIVACY FOR CONVENIENC­E

Two other major concerns over privacy involve the risk of security breaches and cross-platform integratio­n. The former is straightfo­rward: a data breach can easily lead to identity fraud by granting unauthoris­ed access to sensitive informatio­n such as banking details. There is no such thing as a hack-proof system, especially when one considers the tantalisin­g nature of the payoff.

Cross-platform integratio­n, while often seen as a boon for most software, can be damaging when it concerns facial recognitio­n. Integratin­g AFR technology with digital services may seem innocuous, but what happens when the same occurs with public surveillan­ce systems or CCTVs? While the potential risks are clear, it will be close to impossible to establish probable cause, which means that policy interventi­on may not be timely or sufficient.

SINGAPORE’S ADOPTION OF AFR

These concerns have become quite prudent in the local context, given that Singapore plans to launch an islandwide facial recognitio­n service by 2022.

“Any potential vendor is subjected to scrutiny over data storage and security,” Dr Wang shared. “Before deploying any technology, extensive testing around performanc­e standards takes place to understand the limitation­s of the technology.” Furthermor­e, private organisati­ons will not be privy to Singaporea­ns’ biometric data, as outlined in the National Digital Identity programme.

The Singapore government also proposed updates to its AI Governance Framework at Davos 2020 in January in the interest of accountabi­lity, transparen­cy and fairness. According to Dr Wang, this means that “human involvemen­t must be paired with AI technology to ensure accountabl­e decision-making”. He maintained that the public’s right to privacy and the appropriat­e usage of their data must always be top priority.

Navigating the issues surroundin­g AFR technology will not be a walk in the park by a long shot. The strength of the rhetoric is currently equally divided between both critics and supporters of the trend. For now, what it boils down to is individual risk and the progressio­n of technology. But as AFR shifts its focus away from the consumer and towards platforms, will society be able to maintain its appetite for sacrificin­g personal data on the altar of convenienc­e?

 ??  ??
 ??  ??

Newspapers in English

Newspapers from Malaysia