logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo logo
star Bookmark: Tag Tag Tag Tag Tag
Australia

Fears of more social media misinformation, tampering ahead of UK vote

Facebook is opening up a war room to quickly respond to election hoaxes. Twitter is banning political ads. Google plans to crack down on bogus videos on YouTube.

Social media platforms say they are mounting a vigorous campaign against misinformation in the lead up to next month's general election in the United Kingdom. But digital misinformation experts believe British voters remain vulnerable to the same type of misleading ads and phony claims that played a role in the vote to leave the European Union three years ago.

Government inaction on online misinformation and digital ad regulations have added to the pressure internet companies are under as they face growing criticism for amplifying false claims during the run up to the 2016 Brexit referendum and the 2016 election in the US.

Prime Minister Boris Johnson pushed for the snap December 12 election, in which voters will choose 650 representatives to the House of Commons, hoping his Conservative Party will gain enough seats to break a stalemate over his plan to take Britain out of the EU.

And with campaigns barely under way, falsehoods are already spreading online.

A video posted this week on Twitter and Facebook by the Conservative Party contains a misleading edit of a television interview with a senior Labour Party figure. The video had been altered to show the official failing to answer a question about Brexit, when, in fact, he responded quickly.

The chairman of the Conservative Party called the doctored video lighthearted satire, but it's part of a serious problem confronting British voters, according to Will Moy, chief executive at Full Fact, an independent, London-based fact-checking organisation.

"The biggest risk to people in the UK right now is being lied to by their own politicians," said Moy, whose organisation works with Facebook and others as a third-party fact checker. He said laws written decades ago to cover political advertising for print, radio and television can't be applied to the reach and speed of the internet.

Britons may become more divided over Brexit as misinformation and the potential for foreign meddling continues.

Britons may become more divided over Brexit as misinformation and the potential for foreign meddling continues.Credit:AP

Public debate surrounding the 2016 Brexit vote was driven in part by a number of false claims. They included promises that Britain could recoup 350 million pounds per week by leaving the EU; an unfounded claim that a survey later found was believed by nearly half of all Britons.

The threat has grown alongside the influence of social media and the proliferation of online political ads. The proportion of campaign spending on digital advertising has increased from 0.3 per cent in 2011 to 42.8 per cent in 2017, according to the UK's Electoral Commission.

The 2016 US presidential election and the Brexit referendum also highlighted concerns about online foreign interference, after allegations that Russia tried to use social media to divide Americans on hot-button topics like race and religion. A similar tactic may have been used ahead of the Brexit vote: a 2017 study by the University of California Berkeley and Swansea University in Wales found more than 150,000 Twitter accounts with ties to the Kremlin that dispersed messages both supportive and critical of Brexit. Russia has repeatedly denied meddling.

The UK poll will be among the first since the start of Twitter's new policy prohibiting paid political advertisements, which takes effect on November 22. The move was hailed by some as an important step in reducing election misinformation, though critics said it was overly broad and questioned its significance, given Twitter's relatively modest number of political ads.

Loading

"We believe political message reach should be earned and not bought," Twitter CEO Jack Dorsey tweeted.

Twitter's ban stands in stark contrast to Facebook's policy of not fact checking ads from politicians and allowing demonstrably false ads to remain up.

Despite the criticism, Facebook's leaders insist they understand the stakes and take the threat of misinformation seriously.

"We have learned the lessons of 2016, when Russia used Facebook to spread division and misinformation in the US presidential election," Richard Allan, Facebook's vice president of policy solutions, wrote in a piece published last month in The Telegraph.

With 42 million users in Britain, Facebook has the largest social media reach in the UK and has faced the most scrutiny for its role in spreading false information ahead of the Brexit vote. Facebook also owns Instagram and WhatsApp.

Last year, the company began requiring political ads in Britain to carry a disclaimer explaining who paid for it.

Following an 18-month investigation into online privacy and the use of social media to spread disinformation, an influential parliamentary committee in February urged the government to urgently approve new laws to address internet campaign techniques, insisting that democracy itself was under threat.

The Electoral Commission offered its own proposals, including banners on digital political ads clearly identifying their sponsor and increased fines for campaigns that violate the rules.

None of the recommendations have been approved.

"There's been absolutely no response to the recommendations," said Susan Banducci, a political scientist at the University of Exeter who studies the impact that social media plays on elections. "We know there's a problem. You've got to wonder why governments are unwilling to take action."

AP

Most Viewed in Technology

Loading

Themes
ICO