Misinformation thrives on video site popular with far-right
Researchers say nearly 1 in 5 in five videos suggested by YouTube in response to a search for terms related to next week's elections contained misinformation
Your support helps us to tell the story
From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.
At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.
The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.
Your support makes all the difference.Election misinformation is thriving on Rumble, a video sharing platform popular with some conservatives and far-right groups, according to research published Thursday.
Nearly half of the videos suggested by the site in response to searches for common election-related terms contained misinformation, according to the analysis from NewsGuard, a firm that monitors online misinformation.
The percentage was far better at Rumble's much larger rival YouTube, where about 1 in 5 videos were found to contain misinformation after researchers ran the same searches. The search terms included the names of candidates as well as politically sensitive words and phrases such as gun rights, voter fraud and abortion.
The findings illustrate how alternative platforms like Rumble have become hot spots for election-related misinformation as they have increased in popularity. The site is popular with conservatives and some far-right groups critical of content moderation efforts by larger platforms such as YouTube.
Misleading or deceptive claims about voting and elections have proliferated in heading into next week's elections and have been blamed for increasing distrust and polarization.
Some of the videos reviewed by NewsGuard's researchers in October included online shows featuring allies of former President Donald Trump such as Steve Bannon and conspiracy theorists such as Alex Jones. Many videos contained debunked claims about the 2020 election, the Jan. 6, 2021, attack on the U.S. Capitol, the QAnon conspiracy theory, as well as misinformation about voting and the elections.
“Rumble frequently pushes videos from untrustworthy sources that traffic in election misinformation,” NewsGuard's report found.
Messages left with Rumble were not immediately returned Wednesday and Thursday. According to a mission statement on the platform's website, Rumble aims “to restore the internet to its roots by making it free and open once again.”
Rumble said in September it now has 78 million active monthly users around the world, with 63 million in the United States and Canada. The site boasts a long list of podcasts helmed by prominent conservatives such as Don Bongino and Steve Bannon, whose videos have millions of subscribers on Rumble.
The Florida-based platform’s growth has come from users interested in news and politics, as well as younger users in the 18-24 age group, Rumble CEO Chris Pavlovski said in September.
Bannon's show was among the top results when researchers ran a search on the term “voter fraud.” A longtime Trump ally, Bannon was kicked off YouTube last year for repeatedly violating its rules; he was banned from Twitter after calling for Dr. Anthony Fauci, the nation’s leading infectious disease specialist, to be beheaded.
Together, the the misinformation-laden videos turned up by researchers at NewsGuard had been viewed nearly 9 million times so far.
YouTube has been criticized for not doing enough to tackle misinformation on its platform. But the NewsGuard report shows the platform's efforts are making a difference. Researchers said that in addition to suggesting fewer videos containing misinformation, YouTube did not recommend any videos supporting QAnon.
Nevertheless, a report released this fall by New York University faulted Meta, Twitter, TikTok and YouTube for amplifying Trump’s false statements about the 2020 election. The study cited inconsistent rules regarding misinformation as well as poor enforcement.
____
Follow the AP’s coverage of misinformation at https://apnews.com/hub/misinformation.
Follow the AP for full coverage of the 2022 midterm elections at https://apnews.com/hub/2022-midterm-elections and on Twitter at https://twitter.com/ap_politics.
And check out https://apnews.com/hub/explaining-the-elections to learn more about the issues and factors at play in the midterms.