Las Vegas Review-Journal (Sunday)

Can cellphones help predict suicide?

- By Ellen Barry If you are having thoughts of suicide, text the National Suicide Prevention Lifeline at 988 or go to Speakingof­suicide.com/ resources for a list of additional resources.

CAMBRIDGE, Mass. — In March, Katelin Cruz left her latest psychiatri­c hospitaliz­ation with a familiar mix of feelings. She was, on the one hand, relieved to leave the ward, where aides took away her shoelaces and sometimes followed her into the shower to ensure that she would not harm herself.

But her life on the outside was as unsettled as ever, she said in an interview, with a stack of unpaid bills and no permanent home. It was easy to slide back into suicidal thoughts. For fragile patients, the weeks after discharge from a psychiatri­c facility are a notoriousl­y difficult period, with a suicide rate around 15 times the national rate, according to one study.

This time, however, Cruz, 29, left the hospital as part of a vast research project which attempts to use advances in artificial intelligen­ce to do something that has eluded psychiatri­sts for centuries: to predict who is likely to attempt suicide and when that person is likely to attempt it, and then, to intervene.

On her wrist, she wore a Fitbit programmed to track her sleep and physical activity. On her smartphone, an app was collecting data about her moods, her movement and her social interactio­ns. Each device was providing a continuous stream of informatio­n to a team of researcher­s on the 12th floor of the William James Building, which houses Harvard University’s psychology department.

In the field of mental health, few new areas generate as much excitement as machine learning, which uses computer algorithms to better predict human behavior. There is, at the same time, exploding interest in biosensors that can track a person’s mood in real time, factoring in music choices, social media posts, facial expression and vocal expression.

Matthew K. Nock, a Harvard psychologi­st who is one of the nation’s top suicide researcher­s, hopes to knit these technologi­es together into a kind of early-warning system that could be used when an at-risk patient is released from the hospital.

He offers this example of how it could work: The sensor reports that a patient’s sleep is disturbed, she reports a low mood on questionna­ires and GPS shows she is not leaving the house. But an accelerome­ter on her phone shows that she is moving around a lot, suggesting agitation. The algorithm flags the patient. A ping sounds on a dashboard. And, at just the right time, a clinician reaches out with a phone call or a message.

There are plenty of reasons to doubt that an algorithm can ever achieve this level of accuracy. Suicide is such a rare event, even among those at highest risk, that any effort to predict it is bound to result in false positives, forcing interventi­ons on people who may not need them. False negatives could thrust legal responsibi­lity onto clinicians.

Algorithms require granular, longterm data from a large number of people, and it’s nearly impossible to observe large numbers of people who die by suicide. Finally, the data needed for this kind of monitoring raises red flags about invading the privacy of some of society’s most vulnerable people.

Nock is familiar with all these arguments but has persisted, in part out of sheer frustratio­n. “With all due respect to people who’ve been doing this work for decades, for a century, we haven’t learned a great deal about how to identify people at risk and how to intervene,” he said. “The suicide rate now is the same it was literally 100 years ago. So just if we’re being honest, we’re not getting better.”

The gray zone

For psychiatri­sts, few tasks are more nerve-wracking than caring for patients they know to be at risk for suicide while they are at home and unsupervis­ed.

Dr. Karen L. Swartz, a professor of psychiatry at Johns Hopkins University, calls it “the gray zone.” She was fresh out of training when she first wrestled with this problem, caring for a prickly, intelligen­t woman who admitted she had suicidal thoughts, and even alluded to a plan, but dreaded the thought of being hospitaliz­ed.

Swartz turned to the woman’s husband for advice. If you force her into the hospital, he said, she will fire you.

So Swartz decided to take the risk, allowing the woman to remain at home, tweaking her medication­s and waiting. She spent the next weeks on tenterhook­s, and, slowly, the patient improved. “It was one of those things where I just genuinely hoped I was right,” she said. It never gets easier, said Swartz, who now trains young psychiatri­sts: With experience, it only becomes clearer that suicidal thoughts can come and go without warning.

“We are asked to predict something that is highly unpredicta­ble,” she said.

Increasing­ly, health care systems are turning to machine learning to make this call. Algorithms based on vast data sets — drawn from electronic medical records as well as scores of other factors — are used to assign patients a risk score, so that individual­s at exceptiona­lly high risk can be provided with extra attention.

Algorithms have proven more accurate than traditiona­l methods, which, according to a 2017 review of published research, had not improved in 50 years and were only slightly better than chance at predicting an outcome. These methods are already used in some clinical settings. Since 2017, the Department of Veterans Affairs has used an algorithm to flag the 0.1% of veterans at the highest risk for suicide, a few thousand patients in a population of 6 million.

This approach has yielded some success. A study published last year in JAMA Network Open found that veterans enrolled in REACH VET, a program for at-risk patients, were 5% less likely to have a documented suicide attempt, and less likely to be admitted to a psychiatri­c facility or visit the emergency room. But the study found no significan­t change in the rate of suicide.

A fire hose of data

On an August afternoon in the William James building, a lanky data scientist named Adam Bear sat in front of a monitor in Nock’s lab, wearing flip-flops and baggy shorts, staring at the zigzagging graphs of a subject’s stress levels over the course of a week.

When moods are mapped as data, patterns emerge, and it’s Bear’s job to look for them. He spent his summer poring through the days and hours of 571 subjects who, after seeking medical care for suicidal thoughts, agreed to be tracked continuous­ly for six months. While they were being tracked, two died by suicide and between 50 and 100 made attempts.

It is, Nock believes, the largest reservoir of informatio­n ever collected about the daily lives of people struggling with suicidal thoughts.

The team is most interested in the days preceding suicide attempts, which would allow time for interventi­on. Already, some signs have emerged: Although suicidal urges often do not change in the period before an attempt, the ability to resist those urges does seem to diminish. Something simple — sleep deprivatio­n — seems to contribute to that.

Nock has been looking for ways to study these patients since 1994, when he had an experience that shocked him profoundly. During an undergradu­ate internship in the United Kingdom, he was assigned to a locked unit for violent and self-injurious patients. There, he saw things he had never encountere­d: Patients had cuts up and down their arms. One of them pulled out his own eyeball. A young man he befriended, who seemed to be improving, was later found in the Thames.

Another shock came when he began to pepper the clinicians with questions about treating these patients and realized how little they knew: He recalls being told, “We give them some medicine, we talk to them and we hope they get better.”

One reason, he concluded, was that it had never been possible to study a large number of people with suicidal ideation in the same way that we are able to observe patients with heart disease or tuberculos­is. “Psychology hasn’t advanced as much as other sciences because we’ve been largely doing it wrong,” he said. “We haven’t gone out and found some behavior that is important in nature, and gone out and observed it.”

But with the advent of phone-based apps and wearable sensors, he added, “we have data from so many different channels, and we have, increasing­ly, the ability to analyze those data, and observe people as they’re out living their lives.” One dilemma in designing the study was what to do when participan­ts expressed a strong desire to hurt themselves. Nock decided they should intervene.

Telling the truth to a computer

It was around 9 p.m., a few weeks into the six-month study, when the question popped up on Cruz’s phone: “Right now how strong is your desire to kill yourself?”

Without stopping to think, she dragged her finger all the way to the end of the bar: 10. A few seconds later, she was asked to choose between two statements: “I am definitely not going to kill myself today” and “I am definitely going kill myself today.” She scrolled to the second.

Fifteen minutes later, her phone rang. It was a member of the research team calling her. The woman called 911 and kept Cruz on the line until police knocked on her door, and she passed out. Later, when she regained consciousn­ess, a medical team was giving her a sternum rub, a painful procedure used to revive people after overdoses.

Cruz has a pale, seraphic face and a fringe of dark curls. She had been studying for a nursing degree when a cascade of mental health crises sent her life swerving in a different direction. She maintains an A-student’s nerdy interest in science, joking that the rib cage on her T-shirt is “totally anatomical­ly correct.”

Right away, she had been intrigued by the trial, and she responded dutifully six times a day, when the apps on her phone surveyed her about her suicidal thoughts. The pings were intrusive, but also comforting. “It felt like I wasn’t being ignored,” she said. “To have somebody know how I feel, that takes some of the weight off.”

On the night of her attempt, she was alone in a hotel room in Concord, Massachuse­tts. She didn’t have enough money for another night there, and her possession­s were mounded in trash bags on the floor. She was tired, she said, “of feeling like I had nobody and nothing.” Looking back, Cruz said she thought the technology — its anonymity and lack of judgment — made it easier to ask for help.

“I think it’s almost easier to tell the truth to a computer,” she said.

But many in the field are wary of the idea that technology can ever substitute for a clinician’s care. One reason is that patients in a crisis become skilled at deception, said Justin Melnick, 24, a doctoral student who survived a suicide attempt in 2019 and is now an advocate for people with mental illness.

He recalled cutting short telephone conversati­ons with his mother, the person best able to pull him off “the precipice,” and then switching his phone off. “And it was like, OK, that door has been closed,” he said. He described these evasions as “an act of defiance.” Why, he asked, would a person in that frame of mind agree to wear a sensor?

In the end, he said, what helped him turn the corner was people — a support group, which met weekly in a circle of chairs for sessions of dialectica­l behavioral therapy, and a network of friends, family and clinicians who know him well enough to recognize his behavior. When that happens, he said, “we can generally ride that wave together.”

Cruz does not have a network like that. Last month, as temperatur­es in Massachuse­tts were dipping into the 40s, she was living in a tent with her boyfriend, huddling together under a blanket for warmth. In the morning, they waited until Mcdonald’s opened so they could dry out their sweatshirt­s and shoes and charge their devices.

She was faithful about taking her medication­s — five of them — but was scrambling to find a new therapist: The only one in her area who accepts Medicaid has an eight-month waiting list.

 ?? PHOTOS BY KAYANA SZYMCZAK / THE NEW YORK TIMES ?? Katelin Cruz displays a survey question within an app on her smartphone in Ware, Mass. Cruz uses her phone and Fitbit to submit data about her mood and other metrics to Harvard researcher­s studying suicidal tendencies.
PHOTOS BY KAYANA SZYMCZAK / THE NEW YORK TIMES Katelin Cruz displays a survey question within an app on her smartphone in Ware, Mass. Cruz uses her phone and Fitbit to submit data about her mood and other metrics to Harvard researcher­s studying suicidal tendencies.
 ?? ?? Cruz, at a park near her home in Ware, Mass., uses a phone and Fitbit to submit data about her mood and other metrics as part of a Harvard study on how to use artificial intelligen­ce to detect suicidal tendencies.
Cruz, at a park near her home in Ware, Mass., uses a phone and Fitbit to submit data about her mood and other metrics as part of a Harvard study on how to use artificial intelligen­ce to detect suicidal tendencies.

Newspapers in English

Newspapers from United States