San Francisco Chronicle - (Sunday)

Facebook’s tricky role amid suicide threats

- By Natasha Singer

A police officer on the late shift in an Ohio town recently received an unusual call from Facebook.

That day, a woman wrote a Facebook post saying she was walking home and intended to kill herself when she got there, according to a police report on the case. Facebook called to warn the Police Department about the suicide threat.

The officer quickly located

Facebook CEO Mark Zuckerberg

the woman, but she denied having suicidal thoughts, the police report said. Even so, the officer believed she might harm herself and told the woman that she must go to a hospital — either voluntaril­y or in police custody. He ultimately drove her to a hospital for a mental health checkup, an evaluation prompted by Facebook’s interventi­on. (Some details of the case have been withheld for privacy reasons.)

Police stations from Massachuse­tts to Mumbai have received similar alerts from Facebook over the past 18 months as part of what is most likely the world’s largest suicide threat screening and alert program. The social network increased the effort after several people live-streamed their suicides on Facebook Live in early 2017. It now utilizes both algorithms and user reports to flag possible suicide threats.

Facebook’s rise as a global arbiter of mental distress puts the social network in a tricky position at a time when it is

“In the last year, we’ve helped first responders quickly reach around 3,500 people globally.”

under investigat­ion for privacy lapses by regulators in the United States, Canada and the European Union — as well as facing heightened scrutiny for failing to respond quickly to election interferen­ce and ethnic hatred campaigns on its site. Even as CEO Mark Zuckerberg has apologized for improper harvesting of user data, the Menlo Park company grappled recently with fresh revelation­s about special data-sharing deals with tech companies.

The anti-suicide campaign gives Facebook an opportunit­y to frame its work as a good-news story. Suicide is the second-leading cause of death among people ages 15 to 29 worldwide, according to the World Health Organizati­on. Some mental health experts and police officials said Facebook had aided officers in locating and stopping people who were clearly about to harm themselves.

Facebook has algorithms that scan the posts, comments and videos of users in the United States and other countries for indication­s of immediate suicide risk. When a post is flagged, by the technology or a user, it moves to human reviewers at the company, who can call local law enforcemen­t.

“In the last year, we’ve helped first responders quickly reach around 3,500 people globally who needed help,” Zuckerberg wrote in a November post.

But other mental health experts said Facebook’s calls to police could also cause harm — such as unintentio­nally precipitat­ing suicide, compelling nonsuicida­l people to undergo psychiatri­c evaluation­s, or prompting arrests or shootings.

And, they said, it is unclear whether the company’s approach is accurate, effective or safe. Facebook said that, for privacy reasons, it did not track the outcomes of its calls to police. And it has not disclosed exactly how its reviewers decide whether to call emergency responders. Facebook, critics said, has assumed the authority of a public health agency while protecting its process as if it were a corporate secret.

“It’s hard to know what Facebook is actually picking up on, what they are actually acting on, and are they giving the appropriat­e response to the appropriat­e risk,” said Dr. John Torous, director of the digital psychiatry division at Beth Israel Deaconess Medical Center in Boston. “It’s black box medicine.”

Facebook said it worked with suicide prevention experts to develop a comprehens­ive program to quickly connect users in distress with friends and send them contact informatio­n for help lines. It said experts also helped train dedicated Facebook teams, who have experience in law enforcemen­t and crisis response, to review the most urgent cases. Those reviewers contact emergency services only in a minority of cases, when users appear at imminent risk of serious self-harm, the company said.

“While our efforts are not perfect, we have decided to err on the side of providing people who need help with resources as soon as possible,” Emily Cain, a Facebook spokeswoma­n, said in a statement.

In a September post, Facebook described how it had developed a pattern recognitio­n system to automatica­lly score certain user posts and comments for likelihood of suicidal thoughts. The system automatica­lly escalates high-scoring posts, as well as posts submitted by concerned users, to specially trained reviewers.

“Facebook has always been way ahead of the pack,” said John Draper, director of the National Suicide Prevention Lifeline,

If you need help

If you have thoughts of suicide, call the National Suicide Prevention Lifeline at 800-273-8255 or go to www.speakingof­suicide. com/resources.

“not only in suicide prevention, but in taking an extra step toward innovation and engaging us with really intelligen­t and forward-thinking approaches.” (Vibrant Emotional Health, the nonprofit group administer­ing the Lifeline, has advised and received funding from Facebook.)

Facebook said its suicide risk scoring system worked worldwide in English, Spanish, Portuguese and Arabic — except for in the European Union, where data protection laws restrict the collection of personal details like health informatio­n. To opt out, people can stop posting on Facebook or delete their Facebook accounts.

A review of four police reports, obtained by the Times under Freedom of Informatio­n Act requests, suggests that Facebook’s approach has had mixed results. Except for the Ohio case, police department­s redacted the names of the people flagged by Facebook.

In one case in May, a Facebook representa­tive helped police officers in Rock Hill, S.C., locate a man who was streaming a suicide attempt on Facebook Live. On a recording of the call to the police station, the Facebook representa­tive described the background in the video — trees, a street sign — to a police operator and provided the latitude and longitude of the man’s phone.

The Police Department credited Facebook with helping officers track down the man, who tried to flee and was taken to a hospital.

“Two people called the police that night, but they couldn’t tell us where he was,” said Courtney Davis, a Rock Hill police telecommun­ications operator, who fielded the call from Facebook. “Facebook could.”

The Police Department in Mashpee, Mass., had a different experience. Just before 5:16 a.m. Aug. 23, 2017, a Mashpee police dispatcher received a call from a neighborin­g Police Department about a man who was streaming his suicide on Facebook Live. Officers arrived at the man’s home a few minutes later, but by the time they got to him, he no longer had a pulse, according to police records.

At 6:09 a.m., the report said, a Facebook representa­tive called to alert police to the suicide threat.

Scott W. Carline, chief of the Mashpee Police Department, declined to comment. But he said of Facebook, “I’d like to see them improve upon the suicide prevention tools they have in place to identify warning signs that could potentiall­y become fatal.”

Facebook’s Cain said that, in some cases, help unfortunat­ely did not arrive in time. “We really feel for those people and their loved ones when that occurs,” she said.

The fourth case, in May 2017, involved a teenager in Macon, Ga., who was streaming a suicide attempt. Facebook called police after officers had already found the teenager at her home, a spokeswoma­n for the Bibb County sheriff ’s office said. The teen survived the attempt.

Some health researcher­s are also trying to predict suicide risk, but they are using more transparen­t methodolog­y and collecting evidence on the results.

The Department of Veterans Affairs has developed a suicide risk prediction program that uses AI to scan veterans’ medical records for certain medicines and illnesses. If the system identifies a veteran as high risk, the VA offers mental health appointmen­ts and other services. Preliminar­y findings from a VA study reported fewer deaths overall among veterans in the program compared with nonpartici­pating veterans.

In a forthcomin­g article in a Yale law journal, Mason Marks, a health law scholar, argues that Facebook’s suicide risk scoring software, along with its calls to police that may lead to mandatory psychiatri­c evaluation­s, constitute­s the practice of medicine. He says government agencies should regulate the program, requiring Facebook to produce safety and effectiven­ess evidence.

“In this climate in which trust in Facebook is really eroding, it concerns me that Facebook is just saying, ‘Trust us here,’ ” said Marks, a fellow at Yale Law School and New York University School of Law.

Facebook’s Cain disagreed that the program amounted to health screening. “These are complex issues,” she said, “which is why we have been working closely with experts.”

Natasha Singer is a New York Times writer.

 ??  ??
 ?? Jerome Pollos / New York Times 2018 ?? Health law scholar Mason Marks says Facebook is, in effect, practicing medicine.
Jerome Pollos / New York Times 2018 Health law scholar Mason Marks says Facebook is, in effect, practicing medicine.
 ?? Logan R. Cyrus / New York Times 2018 ?? Courtney Davis, a telecommun­ications operator at the Police Department in Rock Hill, S.C., stands with Sgt. Bruce Haire. Facebook called Davis about a man who was live-streaming a suicide attempt.
Logan R. Cyrus / New York Times 2018 Courtney Davis, a telecommun­ications operator at the Police Department in Rock Hill, S.C., stands with Sgt. Bruce Haire. Facebook called Davis about a man who was live-streaming a suicide attempt.

Newspapers in English

Newspapers from United States