Facebook letting fake news spreaders profit, investigators claim
Facebook is allowing users to profit from the spread of potentially dangerous false theories and misinformation about the pandemic and vaccines, including deploying money-raising tools on pages with content flagged up by the social media giant’s own fact-checkers. An investigation has found 430 pages – followed by 45 million people – using Facebook’s tools, including virtual “shops” and fan subscriptions, while spreading false information about Covid-19 or vaccinations.
The findings come despite a promise the platform made last year that no user or company should directly profit from false information about immunization against Covid-19.
Facebook generally does not share this income, but it does occasionally take a cut, and benefits financially from users engaging with content and staying on its services, exposing them to more ads.
The research, by the London-headquartered Bureau of Investigative Journalism, is likely to have uncovered only a tiny snapshot of the vast amount of magnetized misinformation on Facebook related to the pandemic and vaccines.
A Facebook spokesman said the company was investigating the examples brought to its attention, and had “removed a small number of the pages shared with us for violating our policies”.
However, many of the posts identified as misinformation do not violate Facebook rules, the spokesman added, without providing any details.
“Our initial investigation shows a large number of the pages flagged had zero violations against our harmful misinformation policies, and we’d dispute the overall accuracy of the data being provided,” he said.
The pages identified included sites for comedians and religious leaders, social media personalities and traditional media reporters.
There are a large number of alternative health sites, focused on a range of subjects from nutrition to yoga and wellness. Only a minority are clearly focused on the pandemic, or anti-vaccine sentiment. The others are sharing content to broader audiences.
Seven languages are represented – including German, Hebrew, Polish and Spanish – reaching readers around the world.
More than 260 of the pages the bureau identified have posted misinformation about vaccines. The remainder include false information on the pandemic, on vaccines more broadly, or a combination of the two. More than 20 pages identified have gained Facebook’s blue tick signaling authenticity.
For Facebook, offering ways to make money is probably a route to encouraging people to use its platform rather than its competitors’, according to Dr Claire Wardle, executive director of First Draft, a US-based non-profit organization fighting online misinformation, which contributed to the bureau’s research.
However, Facebook can also profit from the popularity of brands and individuals who spread misinformation. It takes a cut of 5% to 30% on its Stars currency, used by fans to tip creators who stream live video.
Facebook also briefly took up to 30% of fees paid by new supporters from January last year, but reversed this in August.
The bureau found two pages using Stars: An0maly and Sid Roth’s It’s Supernatural, a religious site which has blamed the pandemic on abortion and has featured guests describing a dream in which God showed them the virus being created in a Chinese lab. Between them, the pages have reached more than 2.6 million people.
The site run by An0maly, real name AJ Feleski, who describes himself as a “news analyst & hip-hop artist”, is one of the most influential pages sharing misinformation to be identified by the investigation, with more than 1.5 million followers.
A video from last March, in which he questions if the pandemic is “bio-terrorism”, is one of at least three posts on the page that Facebook’s fact-checkers have flagged for containing false or partly false information.
Yet even on Saturday a strap appears under the videos inviting viewers to pay to “Become a supporter” and “Support An0maly and enjoy special benefits”.
Facebook’s policies for creators using monetization tools include rules against misinformation, especially medical misinformation.
In November, Facebook, along with Google and Twitter, agreed a joint statement with the UK government committing to “the principle that no user or company should directly profit from Covid-19 vaccine mis/disinformation. This removes an incentive for this type of content to be promoted, produced and be circulated.”
The bureau’s findings suggest Facebook has breached this agreement, as well as failed to enforce its own policies.
A Facebook spokesman said: “Pages which repeatedly violate our community standards – including those which spread misinformation about Covid-19 and vaccines – are prohibited from monetizing on our platform.
“We take aggressive steps to remove Covid misinformation that leads to imminent physical harm, including false information about approved vaccines.”
The company removed 12 million pieces of Covid misinformation between March and October, and placed fact-check warning labels on 167 million other pieces of content, he added.
Organizations including the UN, the World Health Organization and Unesco said in September that online misinformation “continues to undermine the global response and jeopardizes measures to control the pandemic”.
Some of the pages identified in the investigation also directed their followers to more extreme content that has been largely scrubbed from social media.
Veganize, a Portuguese-language page based in Brazil with 129,000 followers, offers paid supporter subscriptions.
A “pinned post”, which is fixed at the top of the page even as new content is added, carries a link to a collection of files hosted on Google including “Plandemic”, a pair of conspiracy-laden, thoroughly discredited videos that briefly went viral last summer before social networks made strenuous efforts to remove them.
Groups spreading information flagged by fact-checkers as false have also used Facebook to fundraise. The Informed Consent Action Network (ICAN), a US non-profit, is one of the most well-funded organisations in the US opposing vaccinations.
Facebook and YouTube removed pages for Highwire, an online show run by ICAN founder Del Bigtree that made claims repeatedly rated as false by fact-checkers, for which ICAN says it is suing the tech companies.
Yet despite removing the Highwire page, Facebook still allows ICAN to solicit donations from its more than 44,000 followers on a page that has had at least two posts flagged by fact-checkers. According to its page, ICAN has raised almost £24,000 since February 2020.
Facebook must approve organizations signing up to raise funds and vaccine misinformation is explicitly cited as a reason that fundraising may be removed from an organization.
Wardle, from First Draft, believes the money-making systems Facebook offers could encourage people to spread misinformation.
“It is human nature. We know one of the motivations is financial,” she said.
“They have started to believe these things, but when you are in that circle, you also realize there is a way to make money, then you realize that the more you get hits the more money you are making. It’s more than the dopamine hit – it’s dopamine plus dollars.”
Four months after former President Donald Trump was banished from most mainstream social media platforms, he returned to the web last Tuesday with “From the Desk of Donald J. Trump,” essentially a blog for his musings. A week since the unveiling, social media data suggests things are not going well. The ex-president’s blog has drawn a considerably smaller audience than his once-powerful social media accounts.
Florida Gov. Ron DeSantis, continuing his ongoing feud with most of the “corporate media,” on Thursday signed into law a contentious election bill during an event where only Fox News was allowed to observe.
Veteran GOP pollster Frank Luntz warned that former President Trump’s repeated assertions that the 2020 presidential election was rigged against him could hurt Republican efforts to take back the House in 2022. Luntz noted in an interview on the New York Times podcast “Sway” released Thursday that “more than two-thirds of Republicans believe that the election was stolen,” warning that a widespread and unproven belief that there was rampant fraud last November could turn Republicans off from voting in the midterm elections.
The Facebook Oversight Board is a “Supreme Court” for Facebook. On Wednesday, it acted like it — issuing a finely grained ruling that punts the hardest question posed to it back down for Mark Zuckerberg to deal with. The issue before the board, was whether to uphold Facebook’s indefinite ban of Donald Trump’s account following his role in inciting the January 6 riot at the Capitol. It was, by far, the most hotly anticipated decision in the Oversight Board’s young existence. Since the company referred the case to the board on January 21, it received over 9,000 public comments on the matter.