Ofcom: One in three internet users cannot spot false accounts or content

A study by the regulator found that about one in 20 people believe everything they see online.

Josie Clarke
Tuesday 29 March 2022 17:30 EDT
More than a third of internet users are unaware that online content might be false or biased (Dominic Lipinski/PA)
More than a third of internet users are unaware that online content might be false or biased (Dominic Lipinski/PA) (PA Wire)

Your support helps us to tell the story

From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.

At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.

The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.

Your support makes all the difference.

More than a third of internet users are unaware that online content might be false or biased, an Ofcom study has found.

Some 30% of UK adults who use the internet – 14.5 million – are unsure about or do not even consider the truthfulness of the information they see online, the regulator’s annual survey found.

A further 6%, or about one in every 20 internet users, believe everything they see online.

Both adults and children overestimate their ability to spot misinformation, Ofcom found.

Participants were shown social media posts and profiles to determine whether they could verify their authenticity. Although seven in 10 adults (69%) said they were confident in identifying misinformation, only two in 10 (22%) were able to correctly identify the tell-tale signs of a genuine post, without making mistakes.

Ofcom saw a similar pattern among older children aged 12-17, with 74% confident but only 11% able to identify genuine content.

Similarly, about a quarter of adults (24%) and children (27%) who claimed to be confident in spotting misinformation were unable to identify a fake social media profile in practice.

The study also found that 33% of parents of five to seven-year-olds and 60% of parents of eight to 11-year-olds reported their children having a social media profile, despite them being under the minimum age requirement of 13 for most sites.

TikTok, in particular, is growing in popularity, even among the youngest age groups, with 16% of three to four-year-olds and 29% of five to sevens using the platform.

And Ofcom warned that many children could be “tactically” using other accounts or “finstas” – fake Instagrams – to conceal aspects of their online lives from parents.

The study found 64% of eight to 11-year-olds had multiple accounts or profiles, with 46% of these having an account just for their family to see.

More than a third of children (35%) reported engaging in potentially risky behaviours, which could hinder a parent or guardian keeping proper checks on their online use.

A fifth (21%) surfed in “incognito mode” and 19% deleted their browsing history, while 6% circumvented parental controls put in place to stop them visiting certain apps and sites.

Meanwhile, children are seeing less video content from friends online and more from brands, celebrities and influencers, Ofcom said.

Feeds full of “slick professionalised content” seemed to be encouraging a trend towards scrolling instead of sharing, with both adults (88%) and children (91%) three times as likely to watch videos online than to post their own videos.

Ofcom warned that the “sheer volume” of information meant having the critical skills and understanding to decipher fact from fiction had “never been more important”.

Every minute 500 hours of content are uploaded to YouTube, 5,000 videos are viewed on TikTok and 695,000 stories are shared on Instagram.

Four in five adult internet users (81%) want to see tech firms take responsibility for monitoring content on their sites and apps. Two thirds (65%) also want protection against inappropriate or offensive content.

Ofcom chief executive, Melanie Dawes, said: “In a volatile and unpredictable world, it’s essential that everyone has the tools and confidence to separate fact and fiction online – whether it’s about money, health, world events or other people.

“But many adults and children are struggling to spot what might be fake. So we’re calling on tech firms to prioritise rooting out harmful misinformation, before we take on our new role helping to tackle the problem. And we’re offering tips on what to consider when you’re browsing or scrolling.”

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in