The Sunday Guardian

Facebook Live suicides highlight the risks associated with streaming

In recent months, a number of cases involving people live-streaming videos of self-harm or even suicide on the social media platform Facebook have come to light, raising serious concerns about our collective inability to curb harmful content online, write

-

it would do more to try and stop them.

Over the last few weeks, we’ve seen people hurting themselves and others on Facebook—either live or in video posted later,” he wrote. “It’s heartbreak­ing, and I’ve been reflecting on how we can do better for our community.”

In that post, he announced that the company would hire 3,000 more people to its “community operations” team, which already had 4,500 people. They will look over live videos and other content that has been reported, looking not only for people at risk of hurting themselves but banned content like hate speech and child exploitati­on.

The trouble with that is there is simply too much content for any amount of people to look at. Even with huge numbers of people being helped out by even bigger numbers of civilians reporting videos, Facebook simply doesn’t have time to look through videos in realtime and find out ones that show people in danger.

Eventually, some of that work might be taken over by artificial intelligen­ce. Mark Zuckerberg said in a call to investors that because of the size Live video was growing, he expected that it would have to hire more people. Eventually, AI might have to take over, he said—but that will be a matter of years. ver time, the AI t ools will get better,” he said. “Right now, there are certain things that AI can do in terms of understand­ing text and understand­ing what’s in a photo and what’s in a video. That will get better over time. That will take a period of years though to really reach the quality level that we want.

“So for a while our strategy has been to continue building as good of tools as we can. Because no matter how many people we have on the team, we’re never going to be able to look at everything, right? So that’s going to be a big challenge.”

Earlier this year, Facebook added new, less technical features to help people who are at risk of suicide, selfharm or other crises. They primarily exist to put people in contact with groups that can help them out—and the site has said that more technical approaches, like cutting off livestream­s too early, can actually cause problems for the people involved.

If someone sees a friend on Facebook they’re worried about, then they can report the person and the content that worried them. Once that happens, it triggers a range of different features: the person who did the reporting will see tips on how they can help their friend, but the friend themselves will also be given suggestion­s like helplines and tips that can be useful at difficult times.

It introduced the tools with a range of partners, all more experience­d in the work of harm prevention. Those included the Crisis Text Line, the National Eating Disorder Associatio­n and the National Suicide Prevention Lifeline.

Zuckerberg said in May that those features were working —some of the time.

“This is important,” he wrote. “Just last week, we got a report that someone on Live was considerin­g suicide. We immediatel­y reached out to law enforcemen­t, and they were able to prevent him from hurting himself. In other cases, we weren’t so fortunate.”

The site said when it rolled those out that it needed to be careful with cutting of livestream­s, since that might actually put the person at the centre of them at more risk. “Some might say we should cut off the livestream, but what we’ve learned is cutting off the stream too early could remove the opportunit­y for that person to receive help,” Facebook Researcher Jennifer Guadagno told Techcrunch when it was promoting those features. THE INDEPENDEN­T

Earlier this year, Facebook added new, less technical features to help people who are at risk of suicide, self-harm or other crises. They primarily exist to put people in contact with groups that can help them out—and the site has said that more technical approaches, like cutting off livestream­s too early, can actually cause problems for the people involved.

 ??  ?? Facebook added new, less technical features to help people who are at risk of suicide, self-harm or other crises.
Facebook added new, less technical features to help people who are at risk of suicide, self-harm or other crises.
 ??  ??
 ??  ??

Newspapers in English

Newspapers from India