San Antonio Express-News

Using AI to manage employees has risks

- Chris Tomlinson writes commentary about business, economics and policy. twitter.com/cltomlinso­n chris.tomlinson@chron.com

Harried supervisor­s will tell you managing a team is difficult in the best of times when face-to-face interactio­n happens daily and business is good.

Providing constructi­ve feedback to dispersed workers during a once-in-a-lifetime pandemic that has disrupted regular business can feel impossible. Some companies have seized on new tools built on artificial intelligen­ce to monitor their employees.

These nascent productivi­ty monitoring tools, though, come with high risks, according to Serife Tekin, a professor who studies ethics and AI at the University of Texas at San Antonio.

“There’s a big, big hype about artificial intelligen­ce and what it can do,” she told me. “The enthusiasm for these tools has increased since we started the Zoom life with the beginning of the pandemic. But they’re extremely, extremely primitive.”

Most of us have seen chat boxes pop up on websites offering help. In most cases, these are computer programs using AI to answer basic questions. If your question is beyond the AI’S capabiliti­es, the chat is routed to a customer service representa­tive. Some companies have taken AI chatbots to another level, even offering automated psychother­apy. Others offer to embed AI in work computers to detect burnout or other worrisome employee behaviors. An Amazon driver recently quit because he was tired of the company’s AI looking over his shoulder.

“This whole technology is grounded upon this thing called digital phenotypin­g,” Tekin said. “It’s basically inferring people’s mental states based on their digital footprint.”

The AI measures how quickly you speak on the phone or in a Zoom call. It can track how often you visit social media sites or how you respond to text messages. It measures your online behavior to determine your mental health, a correlatio­n that researcher­s discovered more than a decade ago.

“Human-computer interactio­n measures not what you type, but how you type,” Dr. Thomas Insel, a former head of the National Institutes of Health, wrote in a landmark 2008 paper. “Subtle aspects of typing and scrolling, such as the latency be

tween space and character or the interval between scroll and click, are surprising­ly good surrogates for cognitive traits and affective states.”

Insel and other researcher­s are convinced that digital phenotypin­g will revolution­ize mental health care. Hundreds of companies have seized on the research to launch employee monitoring apps, some concentrat­ing on boosting productivi­ty, while others are nothing more than spyware.

AI applicatio­ns can track an employee’s every keystroke or mouse movement. Others use

the computer’s camera to track your eyes. If the AI determines a worker is unusually frustrated or unproducti­ve, it can report the problem to human resources or a supervisor.

The apps can also flag high performers and perhaps give supervisor­s insight into how to boost the productivi­ty of other employees. But like many new technologi­es, these apps often overpromis­e and underperfo­rm.

In its most beneficent use, human resources department­s use AI to detect employee burnout.

“It’s assuming that people respond to stress or burnout similarly,” Tekin said. “Your response to burnout might be turning inward and becoming

even less expressive, exhausted and less communicat­ive. Whereas my response to burnout might be trying to control the situation.”

There are also cultural difference­s in how people cope with stress, depression and anger. The challenge for AI apps is sorting through the variabilit­y of human responses. If developers were sincere, they would admit the complexity of the human mind cannot be understood by a computer program that operates on a broadband internet connection.

“I think it’s extremely problemati­c to be surveillin­g and monitoring employees that way, because this will then imply that it will be in the employee’s records,”

Tekin added. At some point, a behavioral issue could be caused by a medical one, which triggers all kinds of federal and state regulation­s.

Tekin’s biggest concern is that the technology is being deployed more quickly than ethicists or policymake­rs can respond. The Food and Drug Administra­tion has already loosened rules around psychother­apy apps to expand access during the pandemic because so many patients do not want in-person appointmen­ts.

“AI being used to enhance employee-employer relationsh­ips, I think is more acceptable,” Tekin added. “An AI chatbot replacing the kind of work that is expected from the employer

is problemati­c.”

Today’s managers are always looking for ways to work more efficientl­y. Companies are already using AI to screen job applicants, check work performanc­e and gauge employee job satisfacti­on.

Truly empathetic managers, though, will not rely on digital tools to perform the humanistic portion of their jobs. Heartfelt conversati­ons with co-workers are laborious and time-consuming but still the best way to keep top talent on board.

 ??  ??

Newspapers in English

Newspapers from United States