FACEBOOK TESTING WARNS USERS WHO HAVE SEEN ANY ‘HARMFUL EXTREMIST CONTENT’

 6015 views

Facebook is testing a new prompt to warn users if someone has seen harmful extremist content on the social media site.

 

Facebook Announcement

“This test is part of our larger work to assess ways to provide resources and support to people on Facebook who may have engaged with or were exposed to extremist content, or may know someone who is at risk. We are partnering with NGOs and academic experts in this space and hope to have more to share in the future.”

 

The world’s largest social media network has continued to be under pressure from legislators and civil rights groups to resist extremism on its platforms, including U.S. domestic movements associated with the Jan. 6 Capitol riot when groups raising former President Donald Trump tried to stop the U.S. Congress from certifying Joe Biden’s victory in the November election.

Want to Make Free Website Audit Report Click Here: https://www.w3era.com/website-seo-analyzer/

 

Becoming an Extremist?

Screenshot tweeted on Twitter revealed on social media!!

“Are you concerned that someone you know is becoming an extremist?” and another one warned users “you may have been exposed to harmful extremist content recently. “Both included links to “get support.”

 

One asked: “Are you concerned that someone you know is becoming an extremist?”

Another alerted user “You may have been exposed to harmful extremist content recently,”

Both prompts direct users to a page with support linked to extremism.

Facebook stated that tests recognized both users who may have been shown to rule-breaking extremist content and others who had earlier been the subject of implementation on the platform.

Let’s Connect with Us, To Grow Your Business Quick: https://www.w3era.com/contact-us/

It announced the efforts were the role of its commitment to the Christchurch Call to Action, a campaign linking major tech platforms to counter violent extremist content online that was launched following a 2019 attack in New Zealand that was live-streamed on Facebook.

Jess McBeath, the online safety consultant at the UK Safer Internet Centre said-

“We’re seeing an evolving variety of techniques by social media companies to address the seemingly growing problems of mis and disinformation on their platforms,”

In recent years, the company, which has bound its rules against violence and hate groups, said it does remove some content and accounts that violate its rules pro-actively before the matter is seen by users, but that other content may be seen before it is supported against.