Facebook and Google have come under fire from consumer group Which? for failing to take action against online scams.
A new study from Which? claims fraudulent adverts remained online after being reported by users.
The study found 34% of people who reported an advert to Google after falling victim to it said the ad was not taken down, with 26% of Facebook users saying the same.
Both companies said they remove fraudulent adverts and such content was not allowed on their platforms, but Which? said the current reactive approach was not working and called on the Government to include online scams within the scope of its upcoming Online Safety Bill.
The study also found there was a low level of engagement with scam reporting processes on these platforms among the public, with 43% of those who have fallen victim to a scam saying they did not report it to the platform on which they saw it.
Some 31% of those said they did not report a scam advert because they did not think anything would be done about it.
In light of the results Which? has called for a new approach and said online platforms should be given legal responsibility for preventing fake and fraudulent adverts from appearing on their sites, rather than relying on reacting to fake ads once they had caused harm.
Which? consumer rights expert, Adam French, said. “Our latest research has exposed significant flaws with the reactive approach taken by tech giants including Google and Facebook in response to the reporting of fraudulent content – leaving victims worryingly exposed to scams.”
He added: “Which? has launched a free scam alert service to help consumers familiarise themselves with the latest tactics used by fraudsters, but there is no doubt that tech giants, regulators and the government need to go to greater lengths to prevent scams from flourishing.
“Online platforms must be given a legal responsibility to identify, remove and prevent fake and fraudulent content on their sites. The case for including scams in the Online Safety Bill is overwhelming and the Government needs to act now.”
In response to the research, a Facebook company spokesperson said: “Fraudulent activity is not allowed on Facebook and we have taken action on a number of pages reported to us by Which?.
“Our 35,000 strong team of safety and security experts work alongside sophisticated AI to proactively identify and remove this content, and we urge people to report any suspicious activity to us.
“Our teams disable billions of fake accounts every year and we have donated £3 million to Citizens Advice to deliver a UK Scam Action Programme.”
In a statement, Google said: “We’re constantly reviewing ads, sites and accounts to ensure they comply with our policies. As a result of our enforcement actions (proactive and reactive), our team blocked or removed over 3.1 billion ads for violating our policies.
“As part of the various ways we are tackling bad ads, we also encourage people to flag bad actors they’re seeing via our support tool where you can report bad ads directly. It can easily be found on Search when looking for ‘How to report bad ads on Google’ and filling out the necessary information. It is simple for consumers to provide the required information for the Google ads team to act accordingly.
“We take action on potentially bad ads reported to us and these complaints are always manually reviewed.”
“We have strict policies that govern the kinds of ads that we allow to run on our platform. We enforce those policies vigorously, and if we find ads that are in violation we remove them. We utilise a mix of automated systems and human review to enforce our policies.”
Comments: Our rules
We want our comments to be a lively and valuable part of our community - a place where readers can debate and engage with the most important local issues. The ability to comment on our stories is a privilege, not a right, however, and that privilege may be withdrawn if it is abused or misused.
Please report any comments that break our rules.
Read the rules hereComments are closed on this article