Facebook announces Voting Information Center in effort to register 4 million new voters

Facebook have launched a big push to register more peoople for this autumn’s US elections. Among the tools they have created is a ‘Voting Information Center’. From this summer, anyone logging into Facebook, Messenger or Instagram will see a banner advertising the function. Facebook claim they helped 2 million people to register in 2016 and in 2018 and they want to double that number this time.

2_VotingInfoCenter_FBThe Information Center will have information about registering to vote as well as absentee or postal votes, depending on the particular rules of the state they live in.

In addition, Facebook has finalised their opt-out system for political adverts. Users will be able to toggle a switch to block all political and issue based adverts – anything that has a ‘paid for by…’ label. That’s fine, but it is a blunt instrument. There is no ability to choose only to block certain adverts. And it will be interesting to find out (if they will tell us) how many users take up this feature. The good news is that this feature will slowly roll out across other countries that have an advert register.

There are also a couple of small tweaks. The ‘paid for’ disclaimer that indicated a post was an advert used to disappear when an ad was shared. Now that label will stay on the post. Finally, the platform is tracking the amount spent by political contest so that users can identify better what money is being spent where, not just by who. Hopefully that feature will roll across to countries where campaign finance is more tightly regulated as soon as possible.

So, as you might expect, I have a number of concerns about this scheme, even if the overall proposal is very welcome:

  • First, however big and bold they are making it seem, this is still not the grand vision that Facebook has been lacking for so long when it comes to political posts, adverts and electoral interference. Until we know what their long term gameplan is, they will continue to fiddle around the edges.
  • Second, once again we are looking at a big initiative rolled out for a US election. There is absolutely nothing to indicate when such provisions might be made available in the 150+ other countries in which Facebook has a major influence on voters. Yes – the US election is the biggest single contest this year and Facebook is based there. But having a completely America-centric view on things is deeply damaging to the platform’s reputation in many other countries.
  • Third is what is not being said. Facebook is claiming: “By getting clear, accurate and authoritative information to people, we reduce the effectiveness of malicious networks that might try to take advantage of uncertainty and interfere with the election.” My fear is that they will use the existence of the Information Center as an excuse for not acting as they should when leading figures break the platform rules. A month ago President Trump had a post tagged on twitter because it was deemed that he was aiming to spread mistrust in the election system. This was about the only area in which most platforms are prepared to act (although Twitter also censored a post which it claimed was glorifying violence). This week he has again claimed (without justification) that ‘Democrats will stuff ballot boxes with thousands of fake votes’. That, again, is a post aiming to spread mistrust in the election and should have been blocked. But it hasn’t been. If Facebook starts pointing to the information center as the reason they aren’t taking down such posts when they appear on their platform then they will have failed voters rather than served them.

Facebook names their first board members – they have a lot to do

Facebook has picked the first members of its new oversight board which will guide company policy on issues to do with free speech. The line-up so far is impressive (but you would expect that from Facebook). The question is whether this group will be able to wield enough power to change company policy.

Among the four co-chairs of the group is Helle Thorning-Schmidt, the former prime minister of Denmark. She is joined by two US law professors, Jamal Greene and Michael McConnell, and Catalina Botero Marino, a former special rapporteur for freedom of expression at the Organization of American States. Nobel peace laureate Tawakkol Karman and Alan Rusbridger, the former Guardian editor-in-chief, are among the 16 ordinary board members so far selected and the total board will expand to 40 names over time.

“Our roster includes three former judges, six former or current journalists, and other leaders with backgrounds from civil society, academia and public service,” said Thomas Hughes, the director of the oversight board. “They represent a diverse collection of backgrounds and beliefs, but all have a deep commitment to advancing human rights and freedom of expression.”

So what is the board? Well its main aim is to set Facebook policy and act as the final arbiter in disputes about what should and should not be allowed on the platform. This has been an area which has been decidedly lacking until now but has become more and more important.

Take, for example, Facebook’s policy on political speech and adverts. I’ve written a lot about this in the past. I have criticised the company for failing to have a vision as to how it believes politicians (and issue campaigners) should be able to act, a rationale for why they should be treated differently from others and a robust fact-checking system which can guide users to understand why what they are being told might not be true. As a result, Facebook has become out of line with other platforms and often appears to be making up policy on the hoof.

Facebook currently operates a single world-wide policy on political speech. They allow politicians to say what they want. And they allow political adverts to pretty much do the same. In contrast, ordinary advertisers cannot say things that are untrue and even organic posts can be subject to fact-checking. Given the predominance that the platform holds in the marketplace in many countries, this can allow politicians free rein to lie to the electorate with little chance that alternative points of view – or the truth – will get an airing.

Facebook has also failed to take account of the different election laws that apply in countries around the world. Many of these are out-dated, but the platform hasn’t really grasped the chance to work with legislatures to update laws and make sure that Facebook policies in the territory are in compliance.

So will the new board deal with these issues? We will have to wait and see.

Social media and elections – an update

I’ve written a fair amount over the past year about the problems social media companies face with regard to elections and political posts. You can find examples here, here and here.

In short, the argument is that these companies are massively powerful and can have a huge influence on elections. They have chosen different paths with Twitter banning political adverts, Google restricting the targeting that is allowed for such adverts and Facebook regarding the matter as one of free speech and therefore not really making any changes. Plus there is the problem that all companies are working on the basis of a single worldwide policy which pays no regard to the individual laws that might be applicable in the different countries they operate in.

There have been a few updates in recent weeks:

  • Wired has an article which explains the new Facebook Board, a group of people operating at arms length from the platform who will take the final decision on content moderation. As the article points out, they are almost certain to hear a case on false statements in political adverts and, whilst their rulings don’t automatically set a precedent, it may well be that this is the start of a process that leads to a change in policy.
  • Facebook has also updated its political ad library to make it more transparent and given users the option of seeking fewer political adverts in the future.
  • This decision is not without its critics. In the US, many political consultants – used to being able to rely on Facebook’s micro-targeting functions – have suggested that the ability of users to limit the political adverts they see could make campaigning harder. They have produced a report, which it seems Facebook is looking closely at, suggesting ways forward. However, as with so many concerns in the past, this is a purely US campaign industry solution which doesn’t take account of worldwide issues.
  • Twitter came under fire following a BBC exposé which showed that adverts could be targeted at extreme groups such as neo-Nazis. The platform has pledged to ban such adverts in the future.

Facebook, free-speech and political advertising

I previously wrote a long-read about Facebook and the issues they face around elections, political advertising and fact-checking. Since that time, there have been a number of developments (and there will undoubtedly be more) and have received a large number of helpful comments from colleagues around the world. I have therefore chosen to update this piece and republish it here. Many thanks to all those who have assisted me and I welcome further comments. Except where attributed, responsibility for what is written is mine.

 

Facebook’s election-related policies continue to make the news. The issue was highlighted by their decision not to seek to police the truth or otherwise of a Donald Trump advert claiming that potential 2020 opponent Joe Biden used leverage over $1bn of foreign aid to Ukraine to persuade the country to push out the official running an inquiry into his son. 

CNN took the decision not to air the advert which it says has been comprehensively disproved by journalists and organisations such as factcheck.org. However, Facebook, alongside Youtube, Twitter and Fox News, all allowed it to air saying they do not want to get involved in issues of free speech and that the advert does not violate company policies.

“Our approach is grounded in Facebook’s fundamental belief in free expression, respect for the democratic process, and the belief that, in mature democracies with a free press, political speech is arguably the most scrutinized speech there is,” wrote Katie Harbath, Facebook’s head of global elections policy to the Biden campaign.

 

 ELIZABETH WARREN’S FAKE NEWS ADVERT

In response, Elizabeth Warren, another of the leading Democratic candidates produced her own Facebook advert claiming that Mark Zuckerberg and Facebook are openly backing Donald Trump’s re-election. This is not true, of course, as she acknowledges, but she says the advert goes to show how the platform’s decision could be used in the coming contest.

 

FACEBOOK’S DILEMMA

In truth, Facebook’s decision is broadly in line with the situation in the UK. Political, election and candidate adverts (where they are allowed) are subject to different regulation from washing powder or supermarkets. Whereas regular adverts can be judged by Ofcom or the Advertising Standards Authority (depending on the medium), political adverts are subject only to the broader oversight of courts concerned with obscenity, incitement and defamation.

It also highlights the almost impossible task facing the company. If they choose to intervene and judge the truthfulness or otherwise of political statements (either directly or via third parties) then they will be accused by Trump and others of interfering in the right to free speech. If they do not intervene then they will be labelled as allowing lies and distortion to affect the election.

As a colleague puts it:

“The UK tradition about political advertising has been a balance between two considerations. One was that normally the truth-value of a political claim was usually harder to judge than with commercial claims; particularly when the political claim relates to things that will happen in the future. Therefore, claims about how much Labour would put up your taxes if they won, a staple of Conservative election campaigning forever, or the abolition of the NHS, were all debatable. Parties rarely resorted to outright, provable lies about observable facts. But also, political ads were excluded from the ban on ‘knocking copy’ – comparative advertising saying your product or shop is best. But there was a presumption that if a party stepped over an ethical (but not regulatory) line, then another party had the right to a response that was stronger than would normally be permitted in advertising.

That sort of worked, although the partisan press subverted it a bit and there wasn’t a level playing field with campaigning resources (the Conservatives got away with some pretty bad scare stuff in 1924 and 1931, while Labour’s dodgy ‘whose finger on the trigger?’ campaign in 1951 met significant blowback, although it still seemed to be effective).

The problem now is that without a unified media the same process where dodgy claim is met with comparative advertising no longer happens, and with what is left of public service media or media of record afraid to adjudicate on truth value it doesn’t get done. And Facebook’s business model probably makes the exposure to comparative, critical takes even less likely.

And alongside the rise of this, there’s the rise in shameless lying – like terrorism, there’s not much a civilised society can do about bad faith actors in positions of political leadership without subverting its own civilised values.

Facebook is different from the Baldwin era press barons who used their power to pursue their own hobby-horses. It’s power without responsibility for rent – pay enough, or have enough power already, and there aren’t really any barriers. In conditions of extreme upper-end wealth and income concentration, it’s a recipe for abuse.”

THE AD LIBRARY

Facebook’s political advert library is a significant step forward as it allows anyone to see any political or issue based advert that has run on Facebook and will extend to keep a record for seven years. So even if an advert was targeted at a small section of the population, if it was registered as a political advert then it will be in the ad library. The decision to include issue based adverts was highlighted by Facebook’s VP of Policy Solutions, Richard Allan, in a column in the Daily Telegraph.

The problem is that it appears so far only to be operational in 35 countries, a small proportion of the number in which the company operates (albeit a majority by population). In addition, it relies on a complaints system to weed out unregistered political adverts, as exemplified by the first example of a political Facebook advert being banned in the UK – or at least the first example that got widespread publicity. This came from an organisation called the ‘Fair Tax campaign’ and contained the claim that Labour’s tax plans would cost everyone an extra £214 per month. It was taken down by Facebook after a number of complaints to the BBC advert watch initiative headed by journalist Rory Cellan-Jones. 

However, the reason for the advert being taken down is that it didn’t comply with Facebook’s registration and ‘imprint’ rules, not for breaching the rule about false claims. So Facebook have stated that if the advertiser registers it can go back up again (at least until a fact checking organisation takes a look at it and thinks differently). That an advert about tax which prominently attacked a major party and was placed during an election period wasn’t picked up by the company’s algorithm as needing registration indicates that there might be others which have slipped through the net.

 

THE HUMAN RIGHT DIMENSION

In a paper for Chatham House, Kate Jones of Oxford University applies a human rights dimension to the issue of online disinformation and political discourse. She argues that there should only be limited reasons why freedom of speech and the freedom to campaign in elections should be denied. If speech which is claimed to be untrue is banned in mature democracies, will this not be taken as carte blanche for authoritarian governments to restrict opposition voices they do not like?

However Jones also argues that the algorithms used by social media platforms already compromise free speech by restricting, or at least prioritising, what users get to see. And by amplifying more extreme positions, they increase tension and anger in a way which suits their business model but distorts public discourse. She also points out that the right to privacy may be being eroded as personal likes and interests as divulged to a platform are then used to help candidates, parties and advertisers bombard users with information they may not have consented to.

 

WHAT COULD BE DONE?

Could it be different? Certainly. There are many countries where the law states that political adverts and promotions must be truthful in the same way as commercial adverts are. The mnost extreme of which is Singapore where false political statements are taken to be a matter for the police and courts. And whilst Ms Harbath’s comments were reflective only of the situation in the USA, they raise a whole set of questions as regards countries which are not mature democracies, where many people get their news from social media and where the rules of elections are simply different from the USA or UK.

In the past, the platform set different rules for advertisers as opposed to regular posters. They banned adverts containing “deceptive, false or misleading content”, a much stronger restriction than its general rules around non-paid for posts. However, Facebook has now announced that it will allow political adverts to run, regardless of falsehoods they might contain, with the exception, perhaps, of posts that contain links to previously debunked third party content.

 

CLEGG SETS OUT FACEBOOK’S VIEW

Nick Clegg, now Facebook’s VP of Global Affairs and Communications, said in a recent speech:

“…we will not send organic content or ads from politicians to our third-party fact-checking partners for review. However, when a politician shares previously debunked content including links, videos and photos, we plan to demote that content, display related information from fact-checkers, and reject its inclusion in advertisements.”

Clegg goes on to discuss what Facebook refers to as a ‘newsworthiness’ exemption:

“This means that if someone makes a statement or shares a post which breaks our community standards we will still allow it on our platform if we believe the public interest in seeing it outweighs the risk of harm. Today, I announced that from now on we will treat speech from politicians as newsworthy content that should, as a general rule, be seen and heard. However, in keeping with the principle that we apply different standards to content for which we receive payment, this will not apply to ads – if someone chooses to post an ad on Facebook, they must still fall within our Community Standards and our advertising policies.

When we make a determination as to newsworthiness, we evaluate the public interest value of the piece of speech against the risk of harm. When balancing these interests, we take a number of factors into consideration, including country-specific circumstances, like whether there is an election underway or the country is at war; the nature of the speech, including whether it relates to governance or politics; and the political structure of the country, including whether the country has a free press. In evaluating the risk of harm, we will consider the severity of the harm. Content that has the potential to incite violence, for example, may pose a safety risk that outweighs the public interest value. Each of these evaluations will be holistic and comprehensive in nature, and will account for international human rights standards.”

I’m afraid that I don’t quite know what this means. How is Facebook making a judgement about the risk of harm? How will the timing of an election or being at war affect their view? How will the existence of a free press affect the decision and what exactly constitutes a free press – it is often not a binary issue? Katie Harbath told me that:

“In terms of what might cause harm we are looking for things that may jeopardize physical safety. We work with trusted partners across the globe to help us identify if content might lead to real world harm. In terms of if a country has freedom of speech or press we use a few sources, but Freedom House’s rankings are a main one.”

Shortly after Clegg made his speech, Facebook founder Mark Zuckerberg outlined what the company is planning to do for the 2020 US Presidential election. His statement comes after the platform revealed that they had taken action against a number of accounts seeking to disrupt the election originating in Iran and Russia.

Zuckerberg admitted that the company was caught on the back foot in 2016 and needs to do more this time around. He says they will do more to secure the accounts of elected politicians, clearly label posts coming from state media organisations and more clearly label posts deemed false by fact-checkers. They also say they will ban political ads aimed at suppressing turnout. Apparently this will also apply to ads originating from politicians – a provision most likely to hit opposition parties boycotting elections.

The significance of Zuckerberg’s statement is that there appear to be some differences with what Clegg laid out. To what extent will politicians who repeat previously debunked claims have their ads banned or suppressed (or even labelled)?

 

FACT-CHECKERS

Currently, Facebook lists approved third-party fact checkers in 54 countries and one region (the Middle East and North Africa). That is a large proportion of the world but clearly not every country in which the platform operates. The number of organisations ranges from one or two in most nations to six in the USA and eight in India.

Whilst the largest network of fact checkers is the AFP news agency (in 36 nations and one region), in many other cases they are small NGOs. In such cases they have limited time and few resources – The UK fact checking organisation has ten full-time staff members. Even the biggest organisations aren’t able to deal with every claim made during a big election. They can pick and choose some of the biggest or boldest claims and subject them to scrutiny, but the need to be comprehensive and expert often means that the damage has been done and the conversation has moved on long before a third party group manages to convince Facebook to remove a false claim. If this sort of delayed action is to have any meaning then there needs to be some sort of consequence for the advertiser who makes false claims – either financial penalties or some form of suspension or ban.

Even if a third party fact checker has deemed a video or statement to be false, Facebook does not remove it entirely. Harbath told me:

“We reduce the reach (demote it) and add a treatment on it so people can see it has been marked fake. If a politician were to share content that already has that treatment then the label would remain.”

The full policy statement is set out here.

Donie O’Sullivan of CNN makes the point that:

“Facebook’s argument might be more convincing in a world without the platform. The company has helped to create and enhance ideological echo chambers. Some Facebook users only follow and engage with content with which they agree… Given how the Facebook News Feed is determined by an algorithm and the highly targeted nature of Facebook ads, it is entirely possible that a Facebook user could see a false ad from a campaign and not encounter a post that challenges or corrects it.”

PAID FOR VS ORGANIC CONTENT

There is also the issue of the split between paid for advertisements and organic posts by candidates and others. Should there be a difference between how content that Facebook gets paid for are treated and those which are simply ‘community posts’? Harbath told me that content posted by people not directly connected to a candidate or campaign will be subject to fuller moderation than that which comes either from the candidate or a recognised group associated with them.

 

THE BIGGER ISSUE – WHO PAYS?

There is also another, often bigger issue. Who is actually behind the advertising that appears online? Facebook now requires that all political adverts are labelled as such, but whilst some might be pretty obvious – it is badged with the name of the candidate or party – much comes from otherwise unknown individuals or organisations. 

In the UK, as in many other countries, campaigning is allowed by parties and candidates and by third party groups whose spending limits vary according to whether or not they are registered with the electoral commission. But Facebook doesn’t restrict advertising just to permitted participants. It takes the view that responsibility rests with the advertiser to conform with the law. And whilst countries such as India issue formal certificates to candidates which Facebook allows to be uploaded to the site as a guarantee of authenticity, the platform also allows other, non certificate holders, to pay for advertising.

One of the biggest concerns that election-watchers have is that foreign money is influencing elections. Facebook have pretty much shrugged their shoulders at this problem. It is certainly the case that they make clear the country from which the advert is broadcast in the ads library but advertisers can take steps to mask their location and the ad library has different levels of functionality depending on the country.

Facebook told me that whilst they can check for an official identity document from the country concerned and that the funds are paid in local currency and using a local billing address, they have no means of knowing the ultimate origin of the funding. It is also unclear how often such documentation is actually checked. The law in the UK requires that the original source of political funding must be a permissible donor. This can be investigated by the police and electoral commission in cases of doubt. Other countries have similar rules. But Facebook doesn’t even make a statement requiring this from political advertisers.

 

ONE SIZE FITS ALL

As I’ve written before, Facebook is not alone in this dilemma. Nor are they doing nothing. But what they have done is pretty much a one-size-fits-all approach, largely based on an American model. They require political advertisers to register and be identified as described above. They also release details of who has paid what amount on a regular basis.

Talking to Katie Harbath, I was told that the company employs 40 teams and 500 full time staff to cover elections across the world. She told me that for each election they start work about 18 months in advance to try to idientify the threats that might be involved and consider whether to reach out to the country’s election commission to discuss these threats. What they don’t do is seek to work with individual countries to ensure that Facebook’s rules align with the laws and regulations of each country. 

The major problem facing platforms is that election law in virtually every country is not consistent with modern campaigning techniques, especially as regards social media. Richard Allan acknowledged this in his Telegraph article but suggested that Facebook will not be taking the lead in trying to fix it. 

“What constitutes a political ad? Should all online political advertising be recorded in a public archive and should that extend to traditional platforms like billboards, direct mail and newspapers? Should anyone spending over a certain amount on political ads have to declare who their main funders are? Who, if anybody, should decide what politicians can and can’t say in their adverts? These are all questions that can only be properly decided by Parliament and regulators.” 

Facebook could instead be offering to work with parliaments and election commissions on a joint project to help re-shape the law to make it relevant to the current era. Facebook tell me they have signed memorandums of understanding with many electoral commissions in Latin America and in India, which is a significant step in the right direction, but only a small one.

One (hopefully) significant step by Facbook was the hiring of Richard Lappin, formerly the Deputy Head of Elections at ODIHR, the election observation wing of the OSCE. If this signals more of a willingness by the company to engage with election observation missions and with election commissions then this can only be good news.

It’s also worth pointing out that the 18 month lead time is only any use if the election goes ahead as planned. Snap elections cause additional headaches and, whilst I presume Facebook have thought in advance about the situation in the UK in advance of our 2019 general election, I don’t know whether this would be the case in, say, Serbia or Malawi.

Certainly the new Facebook policies were not rolled out in time for the 2019 elections in Tunisia and the EU’s election observation mission to that country has made some strong criticisms of the lack of online regulation in that case, noting (apologies for any translation mistakes from the original French):

“Facebook in Tunisia has not developed transparency tools as in other countries, despite calls from civil society. The “advertising library”, an archiving online advertising created by Facebook, usually shows neither the details of spending performed, nor the advertising history. However, these details are often visible to advertisements targeting voters abroad where self-regulatory measures have sometimes been put in place by Facebook.

The Electoral Law prohibits any financing of campaigns from abroad. As of 13 October, the EOM has observed 87 political advertisements in favor of candidates distributed by Facebook pages managed by administrators whose location is either hidden or located abroad. This lack of transparency undermines verification by the ISIE, civil society or citizens, and deprives them of information on the sources and volume of funding for this online campaign.

The Election Law also provides that candidates for the presidential election must submit to the ISIE a list of all their official accounts on social networks. While the candidate Nabil Karoui provided a list of accounts, candidate Kais Saeed said he is not campaigning online except for a website. For its part, the ISIE did not publish the list of these accounts, and as a result voters were not given the opportunity to identify the pages related to the candidates’ official campaign, nor the identity of their directors.”

 

THE END OF PAID-FOR SOCIAL MEDIA ADVERTISING?

It is worth noting that other social media platforms have taken the decision not to allow political advertising (although they do allow political content). Notably, Twitter has decided to refuse political adverts from 22nd November 2019 and TikTok has said it won’t accept political ads on its platform at all, declaring last week that:

“the nature of paid political ads is not something we believe fits the TikTok platform experience” and that political ads don’t support the platform’s mission “to inspire creativity and build joy.”

Twitter’s decision certainly drew attention as it was made in such aa way as to differentiate the company from Facebook. There is an element of virtue signalling as the platform draws comparatively little revenue from this stream. And it may deflect attention from the platform’s problems with bots, fakery and abuse. 

Twitter CEO Jack Dorsey said:

“This isn’t about free expression. This is about paying for reach. And paying to increase the reach of political speech has significant ramifications that today’s democratic infrastructure may not be prepared to handle.”

In his thread explaining the move, he outlined a series of challenges he says that online platforms face, including:

“machine learning-based optimization of messaging and micro-targeting, unchecked misleading information, and deep fakes. All at increasing velocity, sophistication, and overwhelming scale.”

“It‘s not credible for us to say: “We’re working hard to stop people from gaming our systems to spread misleading info, buuut if someone pays us to target and force people to see their political ad…well…they can say whatever they want!”

One of the criticisms of banning political adverts is that it favours incumbents. The suggestion is that new parties and candidates without a well known face or name will find it impossible to break through if they are not allowed to buy advertising. Dorsey addresses this, saying he has “witnessed many social movements reach massive scale without any political advertising. I trust this will only grow.”

But whilst Twitter may diverge from Facebook on accepting political advertising, they have copied them almost word for word on the issue of whether they will censor comments from politicians. They have stated:

“Everything we do starts with an understanding of our purpose and of the service we provide: a place where people can participate in public conversation and get informed about the world around them.

We assess reported Tweets from world leaders against the Twitter Rules, which are designed to ensure people can participate in the public conversation freely and safely.

We focus on the language of reported Tweets and do not attempt to determine all potential interpretations of the content or its intent.

Presently, direct interactions with fellow public figures, comments on political issues of the day, or foreign policy saber-rattling on economic or military issues are generally not in violation of the Twitter Rules.

However, if a Tweet from a world leader does violate the Twitter Rules but there is a clear public interest value to keeping the Tweet on the service, we may place it behind a notice that provides context about the violation and allows people to click through should they wish to see the content.”

One might also contemplate whether fewer political adverts on platforms which have taken some steps in the right direction (eg Facebook) will simply mean a proliferation in less regulated spaces online. It is unlikely that candidates and parties will simply abandon online advertising if the social networks prohibit them – and of course if one bans ads, the incentives for the others to use ads increases. This is a collective action problem that follows from the inherent lack of effective regulation over the online space.

Even if paid-for adverts are no longer allowed, it seems improbable that Facebook or any other social media platform will ban politicians or the sharing of political content. And that means they will continue to play a significant role in elections, especially for those people who get most of their news from social media, who tend to be younger, less wealthy and have less formal education.

 

FACT CHECKING VS PROFIT

Facebook’s $7bn profits are made possible through the use of as much technology – the agorithms – as possible. Introducing moreregulation or more human beings to fact-check or adjudicate on the validity of political speech only serves to get in the way and cost more. CNN’s Donie O’Sullivan suggests that Facebook might be happy to accept the occasional letter of complaint from Joe Biden in return for not losing Trump’s $20 million of adverts since May 2018, an amount that can only rise as the 2020 election approaches. Others believe that the 2020 US elections will be the last in which the platform accepts political advertising. Inevitably, increased transparency measures will reduce candidates’ incentives to use adverts, which in turn will hit Facebook’s profits. In Twitter’s case this might have been judged to be a price worth paying given the (relatively) small income and PR coup they gained from their decision. But for Facebook there is a much larger sum at play.

 

PROPOSALS

Technology platforms have a lot on their plate. Deepfakes, disruptive bad actors and co-ordinated inauthentic behaviour were all raised with me as challenges to be faced. But it remains a disappointment that the most high profile of these companies is still leaving the door so wide for illegal manipulation and are not doing more to recognise that many countries in the world operate elections on a model (and electoral regulation) far different from the US system. 

The Oxford Technology and Elections Commission has reported and made a series of recommendations for actions the UK should be taking to reflect the impact that online campaigning, including social media, has on elections in the UK. These range from an industry implementation of a library system, improved due diligence and imprinting by the parties and verification of social media accounts by the Electoral Commission. Crucially they also recommend that existing financial reporting rules need to be extended to cover all online campaigning. However they do not explicitly require that those who place online adverts declare the origin of the money they spend which I think is a significant shortcoming.

Here are a few things I think platforms could do:

  1. Respect the laws of the country Facebook is operating in rather than seeking to impose a single set of (largely Californian) values world-wide. Work with parliaments and election commissions to help to design political advertising rules for the platform that align with the individual laws of that country, and offer to work with these same bodies to modernise election law where it is deficient;
  2. Roll out the advert library and fact-checking system to every country in which the platform operates and ensure that fact-checkers are funded sufficiently to enable them to do a good job;
  3. As part of the registration process, require all political advertisers to state that they are the original source of the money paying for the adverts or that they have raised it from permissible sources. Use the platform’s own technology to investigate whether this is the case as much as possible;
  4. In cases where national or platform rules are broken, pledge to share all information with election regulators and law enforcement bodies in the country concerned to enable investigation and prosecution.

Ultimately, these suggestions may conflict with free speech ideals and some may worry that protesters in authoritarian states will have their details handed over to the authorities. I acknowledge that these are both legitimate concerns. However I would suggest that this only highlights the need to change the laws in those countries to make them into more mature democracies which embrace legitimate protest and free speech. But there cannot be one rule for us and another for ‘them’. If we want Facebook to take action to ensure that the Russians cannot interfere in elections in the US or UK, then we have to accept that the Russian election authorities will want to enforce the laws that exist in that country too. Facebook could consider prominent links to respected election observer reports (such as by OSCE/ODIHR , the OAS or EU) which highlight shortcomings in a country’s election structures and media freedom.

 

Reading List – Monday 28th October 2019

Three articles related to social media and political advertising today. As a reminder, you can read my own investigations into the matter here and here.

 

A Buzzfeed investigation has discovered that more than 160 political adverts have been removed from Facebook in the first half of October alone for breaching site rules, even though the platform claims that it does not censor political adverts.

“None of these political ads were rejected on the basis of being deemed false by our fact-checkers,” a Facebook spokesperson told BuzzFeed News. “They were removed for violating one or more of Facebook’s other advertising policies, such as our policy against using fake buttons in ads.” The Trump ad, for example, was taken down for violating Facebook’s rules against profanity in advertisements, the company spokesperson explained.

 

Sue Halpern in the New Yorker reviews Mark Zuckerberg’s appearance in front of Congress last week and discusses why Facebook should not be seen as just another boradcaster, nor as a neutral platform. Nick Clegg also gets a namecheck.

 

Finally, the Guardian reports that Facebook’s independent fact-checkers only learned that they were meant to be checking advertisements fro accuracy from Mark Zuckerberg’s statements to congress. However this only applies to commercial adverts and not to political or issue-based content.

Facebook’s dilemma over free speech and what they could be doing to keep elections fair

UPDATE: This long-read has been updated in the light of new developments and comments from a wide range of colleagues around the world. For the newer version, please see here.

 

Facebook’s election-related struggles are continuing to make the news. This time it is their decision not to seek to police the truth or otherwise of a Donald Trump advert claiming that potential 2020 opponent Joe Biden used leverage over $1bn of foreign aid to Ukraine to persuade the country to push out the official running an inquiry into his son. 

CNN has already taken the decision not to air the advert which is says has been comprehensively disproved by journalists and organisations such as factcheck.org. However, Facebook, alongside Youtube, Twitter and Fox News, have all run it saying they do not want to get involved in issues of free speech and that the advert does not violate company policies.

“Our approach is grounded in Facebook’s fundamental belief in free expression, respect for the democratic process, and the belief that, in mature democracies with a free press, political speech is arguably the most scrutinized speech there is,” wrote Katie Harbath, Facebook’s head of global elections policy to the Biden campaign. 

Elizabeth Warren’s fake news advert

In response, Elizabeth Warren, another of the leading Democratic candidates has produced her own Facebook advert claiming that Mark Zuckerberg and Facebook are openly backing Donald Trump’s re-election. This is not true, of course, as she acknowledges, but she says the advert goes to show how the platform’s decision could be used in the coming contest.

Facebook’s dilemma

In truth, the Facebook decision is broadly in line with the situation in the UK. Political, election and candidate adverts (where they are allowed) are subject to different regulation from washing powder or supermarkets. Whereas regular adverts can be judged by Ofcom or the Advertising Standards Authority (depending on the medium), political adverts are subject only to the broader oversight of courts concerned with obscenity, incitement and defamation.

It also highlights the almost impossible task facing the company. If they choose to intervene and judge the truthfulness or otherwise of political statements (either directly or via third parties) then they will be accused by Trump and others of interfering in the right to free speech. If they do not intervene then they will be labelled as allowing lies and distortion to affect the election.

As a colleague puts it:

“The UK tradition about political advertising has traditionally been a balance between two considerations. One was that normally the truth-value of a political claim was usually harder to judge than with commercial claims; particularly when the political claim relates to things that will happen in the future. Therefore, claims about how much Labour would put up your taxes if they won, a staple of Conservative election campaigning forever, or the abolition of the NHS, were all debatable. Parties rarely resorted to outright, provable lies about observable facts. But also, political ads were excluded from the ban on ‘knocking copy’ – comparative advertising saying your product or shop is best. But there was a presumption that if a party stepped over an ethical (but not regulatory) line, then another party had the right to a response that was stronger than would normally be permitted in advertising.

That sort of worked, although the partisan press subverted it a bit and there wasn’t a level playing field with campaigning resources (the Conservatives got away with some pretty bad scare stuff in 1924 and 1931, while Labour’s dodgy ‘whose finger on the trigger?’ campaign in 1951 met significant blowback, although it still seemed to be effective).

The problem now is that without a unified media the same process where dodgy claim is met with comparative advertising no longer happens, and with what is left of public service media or media of record afraid to adjudicate on truth value it doesn’t get done. And Facebook’s business model probably makes the exposure to comparative, critical takes even less likely.

And alongside the rise of this, there’s the rise in shameless lying – like terrorism, there’s not much a civilised society can do about bad faith actors in positions of political leadership without subverting its own civilised values.

Facebook is different from the Baldwin era press barons who used their power to pursue their own hobby-horses. It’s power without responsibility for rent – pay enough, or have enough power already, and there aren’t really any barriers. In conditions of extreme upper-end wealth and income concentration, it’s a recipe for abuse.”

What could be done?

Could it be different? Certainly. There are many countries where the law states that political adverts and promotions must be truthful in the same way as commercial adverts are. And whilst Ms Harbath’s comments were reflective only of the situation in the USA, they raise a whole set of questions as regards countries which are not mature democracies, where most people get their news from social media and where the rules of elections are simply different from the USA or UK.

In the past, the platform set different rules for advertisers as opposed to regular posters. They banned adverts containing “deceptive, false or misleading content”, a much stronger restriction than its general rules around non-paid for posts. However, Facebook has now announced that it will allow political adverts to run, regardless of falsehoods they might contain, with the exception, perhaps, of posts that contain links to previously debunked third party content.

Clegg sets out Facebook’s view

Nick Clegg, now Facebook’s VP of Global Affairs and Communications, said in a recent speech:

“…we will not send organic content or ads from politicians to our third-party fact-checking partners for review. However, when a politician shares previously debunked content including links, videos and photos, we plan to demote that content, display related information from fact-checkers, and reject its inclusion in advertisements.”

Clegg goes on to discuss what Facebook refers to as a ‘newsworthiness’ exemption:

“This means that if someone makes a statement or shares a post which breaks our community standards we will still allow it on our platform if we believe the public interest in seeing it outweighs the risk of harm. Today, I announced that from now on we will treat speech from politicians as newsworthy content that should, as a general rule, be seen and heard. However, in keeping with the principle that we apply different standards to content for which we receive payment, this will not apply to ads – if someone chooses to post an ad on Facebook, they must still fall within our Community Standards and our advertising policies.

When we make a determination as to newsworthiness, we evaluate the public interest value of the piece of speech against the risk of harm. When balancing these interests, we take a number of factors into consideration, including country-specific circumstances, like whether there is an election underway or the country is at war; the nature of the speech, including whether it relates to governance or politics; and the political structure of the country, including whether the country has a free press. In evaluating the risk of harm, we will consider the severity of the harm. Content that has the potential to incite violence, for example, may pose a safety risk that outweighs the public interest value. Each of these evaluations will be holistic and comprehensive in nature, and will account for international human rights standards.”

I’m afraid that I don’t quite know what this means. How is Facebook making a judgement about the risk of harm? How will the timing of an election or being at war affect their view? How will the existence of a free press affect the decision and what exactly constitutes a free press – it is often not a binary issue? Katie Harbath told me that:

“In terms of what might cause harm we are looking for things that may jeopardize physical safety. We work with trusted partners across the globe to help us identify if content might lead to real world harm. In terms of if a country has freedom of speech or press we use a few sources, but Freedom House’s rankings are a main one.”

Even if third party fact-checkers are used, they are generally small NGOs. They have limited time and few resources. They certainly aren’t able to deal with every claim made in a political Facebook advert during a big election such as a general election in the UK. They can pick and choose some of the biggest or boldest claims and subject them to scrutiny, but the need to be comprehensive and expert often means that the damage has been done and the conversation has moved on long before a third party group manages to convince Facebook to remove a false claim. If this sort of delayed action is to have any meaning then there needs to be some sort of consequence for the advertiser who makes false claims – either financial penalties or some form of suspension or ban.

Even if a third party fact checker has deemed a video or statement to be false, Facebook does not remove it entirely. Harbath told me:

“We reduce the reach (demote it) and add a treatment on it so people can see it has been marked fake. If a politician were to share content that already has that treatment then the label would remain.”

Donie O’Sullivan of CNN makes the point that:

“Facebook’s argument might be more convincing in a world without the platform. The company has helped to create and enhance ideological echo chambers. Some Facebook users only follow and engage with content with which they agree… Given how the Facebook News Feed is determined by an algorithm and the highly targeted nature of Facebook ads, it is entirely possible that a Facebook user could see a false ad from a campaign and not encounter a post that challenges or corrects it.”

Paid for vs organic content

And then there is the split between paid for advertisements and organic posts by candidates and others. Should there be a difference between how content that Facebook gets paid for are treated and those which are simply ‘community posts’. Harbath told me that content posted by people not directly connected to a candidate or campaign will be subject to fuller moderation than that which comes either from the candidate or a recognised group associated with them.

The bigger issue – who pays?

There is also another, often bigger issue. Who is actually behind the advertising that appears online? Particularly in countries other than the USA. Facebook now requires that all political adverts are labelled as such, but whilst some might be pretty obvious – it is badged with the name of the candidate or party – much comes from otherwise unknown individuals or organisations. 

In the UK, as in many other countries, campaigning is allowed by parties and candidates and by third party groups whose spending limits vary according to whether or not they are registered with the electoral commission. But Facebook doesn’t restrict advertising just to permitted participants. It takes the view that responsibility rests with the advertiser to conform with the law. And whilst countries such as India issue formal certificates to candidates which Facebook allows to be uploaded to the site as a guarantee of authenticity, the platform also allows other, non certificate holders, to pay for advertising.

One of the biggest concerns that election-watchers have is that foreign money is influencing elections. Facebook have pretty much shrugged their shoulders at this problem. They told me that whilst they can check for an official identity document from the country concerned and that the funds are paid in local currency and using a local billing address, they have no means of knowing the ultimate origin of the funding. It is also unclear how often such documentation is actually checked. The law in the UK requires that the original source of political funding must be a permissible donor. This can be investigated by the police and electoral commission in cases of doubt. Other countries have similar rules. But Facebook doesn’t even make a statement requiring this from political adverisers.

One size fits all

As I’ve written before, Facebook is not alone in this dilemma. Nor are they doing nothing. But what they have done is pretty much a one-size-fits-all approach, largely based on an American model. They require political advertisers to register and be identified as described above. They also release details of who has paid what amount on a regular basis.

Talking to Katie Harbath, I was told that the company employs 40 teams and 500 full time staff to cover elections across the world. She told me that for each election they start work about 18 months in advance to try to idientify the threats that might be involved and consider whether to reach out to the country’s election commission to discuss these threats. What they don’t do is seek to work with individual countries to ensure that Facebook’s rules align with the laws and regulations of each country. In many cases those laws are pretty out of date and were written for a pre-internet age, but they are still the law. Facebook could also be offering to work with parliaments and election commissions on a joint project to help re-shape the law to make it relevant to the current era. Whilst none of this appears to be happening yet, Facebook tell me they have signed memorandums of understanding with many electoral commissions in Latin America and in India, which is a significant step in the right direction. 

It’s also worth pointing out that the 18 month lead time is only any use if the election goes ahead as planned. Snap elections cause additional headaches and, whilst I presume Facebook might have thought in advance about the situation in the UK, I don’t know whether this would be the case in, say, Serbia or Malawi.

Fact checking vs profit

Facebook’s $7bn profits are made possible through the use of as much technology – the agorithms – as possible. Introducing more human beings to fact-check or adjudicate on the validity of political speech only serves to get in the way and cost more. CNN’s Donie O’Sullivan suggests that Facebook might be happy to accept the occasional letter of complaint from Joe Biden in return for not losing Trump’s $20 million of adverts since May 2018, an amount that can only rise as the election approaches. Others believe that the 2020 US elections will be the last in which the platform accepts political advertising.

The end of paid-for social media advertising?

It is worth noting that other social media platforms have taken the decision not to allow political advertising (although they do allow political content). Notably, TikTok has said it won’t accept political ads on its platform, declaring last week that “the nature of paid political ads is not something we believe fits the TikTok platform experience” and that political ads don’t support the platform’s mission “to inspire creativity and build joy.”

Even if paid-for adverts are no longer allowed, it seems improbable that Facebook or any other social media platform will ban politicians or the sharing of political content. And that means they will continue to play a significant role in elections, especially for those people who get most of their news from sociel media.

Proposals

Technology platforms have a lot on their plate. Deepfakes, disruptive bad actors and co-ordinated inauthentic behaviour were all raised with me as challenges to be faced. But it remains a disappointment that the most high profile of these companies is still leaving the door so wide for illegal manipulation and are not doing more to recognise that many countries in the world operate elections on a model (and electoral regulation) far different from the US system. Here are a few things I think they could do while still adhering to the general principles of free speech:

  1. Respect the laws of the country Facebook is operating in rather than seeking to impose a single set of (largely Californian) values world-wide. Work with parliaments and election commissions to help to design political advertising rules for the platform that align with the individual laws of that country, and offer to work with these same bodies to modernise election law where it is deficient;
  2. Require all political advertisers to state that they are the original source of the money paying for the adverts or that they have raised it from permissible sources. Use the platform’s own technology to investigate whether this is the case as much as possible;
  3. In cases where national or platform rules are broken, have in place a system for financial penalties and/or platform bans and pledge to share all information with law enforcement bodies in the country concerned to enable investigation and prosecution.
  4. If reliance is to be placed on fact checking NGOs to counter the most egregious cases, then Facebook should be helping to set up and fund a network of such groups across all the democracies where the platform operates.

Ultimately, these suggestions may conflict with free speech ideals and some may worry that protesters in authoritarian states will have their details handed over to the authorities. I acknowledge that these are both legitimate concerns. However I would suggest that this only highlights the need to change the laws in those countries to make them into more mature democracies which embrace legitimate protest and free speech. But there cannot be one rule for us and another for ‘them’. If we want Facebook to take action to ensure that the Russians cannot interfere in elections in the US or UK, then we have to accept that the Russian election authorities will want to enforce the laws that exist in that country too. Facebook could consider prominent links to respected election observer reports (such as by OSCE/ODIHR , the OAS or EU) which highlight shortcomings in a country’s election structures and media freedom.