Content Moderation Case Study: Facebook's Moderation Of Terrorist Content Results In The Removal Of Journalists' And Activists' Accounts (June 2020)
Furnished content.
Summary: In almost every country in which it offers its service, Facebook has been asked -- sometimes via direct regulation -- to limit the spread of "terrorist" content.But moderating this content has proven difficult. It appears the more aggressively Facebook approaches the problem, the more collateral damage it causes to journalists, activists, and others studying and reporting on terrorist activity.
Because documenting and reporting on terrorist activity necessitates posting of content considered to be "extremist," journalists and activists are being swept up in Facebook's attempts to purge its website of content considered to be a violation of terms of service, if not actually illegal.The same thing happened in another country frequently targeted by terrorist attacks.
In the space of one day, more than 50 Palestinian journalists and activists had their profile pages deleted by Facebook, alongside a notification saying their pages had been deactivated for "not following our Community Standards.""We have already reviewed this decision and it can't be reversed," the message continued, prompting users to read more about Facebook's Community Standards.There appears to be no easy solution to Facebook's over-moderation of terrorist content. With algorithms doing most of the work, it's left up to human moderators to judge the context of the posts to see if they're glorifying terrorists or simply providing information about terrorist activities.Decisions to be made by Facebook:
edit: Policy/auto___content_moderation_case_study__facebook_s_moderation_of_terrorist_content_results_in_the_removal_of_journalists__and_activists__accounts__june_2020_.wikieditish...