Wednesday, August 19, 2020

Facebook removes nearly 800 QAnon-related groups, pages, hashtags, and ads
By Meira Gebel August 19, 2020 

Facebook took down nearly 800 groups associated with the far-right conspiracy theory group QAnon on Wednesday, as well as more than 1,500 advertisements and 100 pages tied to the group in a move to restrict “violent acts.”

In a blog post, Facebook said the action is part of a broader “Dangerous Individuals and Organizations” policy measure to remove and restrict content that has led to real-world violence. The policy will also impact militia groups and political protest organizations like Antifa.

“While we will allow people to post content that supports these movements and groups, so long as they do not otherwise violate our content policies, we will restrict their ability to organize on our platform,” the company said.

QAnon supporters believe in a widely disproven “deep state” conspiracy that President Donald Trump is working to eradicate pedophilia and Satanism throughout Washington D.C. The conspiracy theorists have recently latched onto the COVID-19 public health crisis, calling it a “bioweapon.”

QAnon theories hit the mainstream after the controversy surrounding #Pizzagate, in which a man brought a gun to a pizzeria, claiming he would find victims of child abuse. The group has also been linked to dozens of other violent incidents that stem from baseless theories shared on private Facebook groups and message boards.

Facebook took action against QAnon earlier this month, when it yanked down an influential group with more than 200,000 members, but Wednesday’s move is perhaps the social media giant’s most substantial move yet.

The company said it will limit QAnon content from appearing in its recommendations tab, reduce its content in search results, and prohibit QAnon-related accounts and groups from monetizing content, selling merchandise, fundraising, and purchasing advertising on both Facebook and Instagram. The company plans to continue to investigate just how QAnon operates on its platform, by observing “specific terminology and symbolism used by supporters to identify the language used by these groups and movements indicating violence and take action accordingly.”


In recent months, other social media sites like Twitter and TikTok have banned and disabled popular QAnon hashtags and accounts for inauthentic, coordinated behavior and for spreading disinformation.

However, do not expect QAnon to disappear quietly: Experts have called QAnon members “really good at adapting” to online ecosystems, and several QAnon supporters have won primaries for public office on platforms that represent the conspiracy theories shared within the group.



No comments: