e dot dot dot
home << Policy << auto content moderation case study facebook s moderation of terrorist content results in the removal of journalists and activists accounts june 2020

Thu, 29 Oct 2020

Content Moderation Case Study: Facebook's Moderation Of Terrorist Content Results In The Removal Of Journalists' And Activists' Accounts (June 2020)
Furnished content.


Summary: In almost every country in which it offers its service, Facebook has been asked -- sometimes via direct regulation -- to limit the spread of "terrorist" content.But moderating this content has proven difficult. It appears the more aggressively Facebook approaches the problem, the more collateral damage it causes to journalists, activists, and others studying and reporting on terrorist activity.

Because documenting and reporting on terrorist activity necessitates posting of content considered to be "extremist," journalists and activists are being swept up in Facebook's attempts to purge its website of content considered to be a violation of terms of service, if not actually illegal.
The same thing happened in another country frequently targeted by terrorist attacks.
In the space of one day, more than 50 Palestinian journalists and activists had their profile pages deleted by Facebook, alongside a notification saying their pages had been deactivated for "not following our Community Standards.""We have already reviewed this decision and it can't be reversed," the message continued, prompting users to read more about Facebook's Community Standards.
There appears to be no easy solution to Facebook's over-moderation of terrorist content. With algorithms doing most of the work, it's left up to human moderators to judge the context of the posts to see if they're glorifying terrorists or simply providing information about terrorist activities.Decisions to be made by Facebook:Questions and policy implications to consider:Resolution: Facebook continues to struggle to eliminate terrorist-linked content from its platform. It appears to have no plan in place to reduce the collateral damage caused by its less-than-nuanced approach to a problem that appears -- at least at this point -- unsolvable. In fact, its own algorithms have generated extremist content by auto-generating "year in review" videos utilizing "terrorist" content uploaded by users, but apparently never removed by Facebook.Facebook's ongoing efforts with the Global Internet Forum to Counter Terrorism (GIFCT) probably aren't going to limit the collateral damage to activists and journalists. Hashes of content designated "extremist" are uploaded to GIFCT's database, making it easier for algorithmic moderation to detect and remove unwanted content. But utilizing hashes and automatic moderation won't solve the problem facing Facebook and others: the moderation of extremist content uploaded by extremists and similar content uploaded by users who are reporting on extremist activity. The company continues to address the issue, but it seems likely this collateral damage will continue until more nuanced moderation options are created and put in place.

Read more here


edit: Policy/auto___content_moderation_case_study__facebook_s_moderation_of_terrorist_content_results_in_the_removal_of_journalists__and_activists__accounts__june_2020_.wikieditish...

Password:
Title:
Body:
Link | Image | Paragraph | BR | Return | Create Amazon link | Technorati tag
Technorati tag?:
Delete this item?:
Treat as new?:
home << Policy << auto content moderation case study facebook s moderation of terrorist content results in the removal of journalists and activists accounts june 2020