How Could Deepfakes Impact the 2020 US Elections?

Nick Diakopoulos
11 min readJun 12, 2019

--

By: Nicholas Diakopoulos (Northwestern) and Deborah Johnson (U. Virginia)

New technologies used to produce deepfakes are rapidly advancing and becoming more accessible, allowing users to make compelling video and audio clips of individuals doing and saying things they never did or said. Users can, for instance, synthesize an individual’s voice, swap one person’s face onto another person’s body in a video, or alter a video interviewee’s words merely by re-writing the transcript. Recorded audio-visual media is becoming more and more malleable, facilitating an ease of editing almost analogous to text.

The technology offers a host of potential benefits in entertainment and education from multi-lingual advertising campaigns to museums bringing dead artists back to life. But it can also challenge aural and visual authenticity and enable the production of disinformation by bad actors. Deepfakes have the potential to wreak havoc in contexts, such as news, where audio or video are treated as a form of evidence that something actually happened. So-called “cheapfakes”, such as the widely circulated clip of House Speaker Nancy Pelosi, have already demonstrated the potential for low-tech manipulated video to find a ready audience. The more advanced technology creates a whole new level of speed, scale, and potential for personalization of such disinformation.

The goal of this article is to stimulate reflection on the ethics and governance of these emerging technologies. Specifically we’re focused on the use of these technologies in the context of the 2020 US election and seek to encourage debate about potential responses by various stakeholders: What should social media platforms, journalists, technology developers, and policy makers do to ensure that the outcome of democratic processes aren’t negatively impacted by deepfakes?

To do this we have developed a set of scenarios that describe an array of possible uses (arguably, misuses or unethical uses) of deepfake technology in the 2020 elections. These speculative fictions explore how the current state-of-the-art technology could be deployed by actors with various motivations to impact election outcomes. The scenarios describe a rich and complex constellation of how the technology might interact with human behavior. The situations generate a number of ethical issues and point to dimensions of elections in which norms, policy, regulation, or technical intervention might be needed or helpful to protect the integrity of the 2020 election.

The set of scenarios purposely include a variety of different actors (e.g. candidate or campaign staff, versus external entities like PACs or foreign governments), motivations for those actors (e.g. support a candidate, hurt a competitor, undermine the process), modalities of media (e.g. audio, video, image), phases (e.g. early vs. late), channels for distribution (e.g. social media, podcasts, chat apps), and mechanisms for influencing voters (e.g. discrediting a candidate’s reputation by association, exaggerating a candidate’s views, suggesting a candidate engaged in corruption, providing evidence of a candidate’s hypocrisy, inciting a campaign’s base, intimidating voters, undermining or attacking the election process, and more).

From an ethical perspective, all of the situations described in the scenarios are problematic insofar as they involve deception. However, they vary in the actors that produce and/or distribute the deepfake, the kind of damage or attack they attempt to do, and how the deception can be counteracted. One overarching question is how these variations affect the ethical nature of the situation.

We developed the scenarios with an eye to making them plausible — describing what you might reasonably believe could happen — rather than merely possible. The challenge for you, then, is to consider what might make the plausible less probable. What can be done now — in the way of establishing norms, rules, policies, etc. — to avoid the worst outcomes (or at least make them less likely)? The scenarios, and brief reflections on each are below. We’d love to get your feedback as to whether we have achieved our goals.

Scenario #1

A small veterans’ organization would like to see Seth Moulton win the democratic primary because although he is not the only candidate with military experience, he is the only one with significant combat experience and the only one making veterans issues a central component of his campaign. The group becomes a Political Action Committee (PAC) and raises funds to make a promotional video for Moulton. The video consists of a combination of video clips with voice over that valorize his bravery. One of the clips is a synthesized depiction of an incident in Iraq when Moulton heroically saved the lives of several members of his platoon. The video is posted on YouTube without any indication that one portion is synthesized. Thousands view the video; hundreds make comments. From the comments, it is apparent that some viewers believe the video is real footage taken by a reporter present at the event. Other comments include complaints from soldiers in Moulton’s platoon who claim that the depiction is an exaggeration of what happened. Because of the comments the video has mixed effects: some are convinced that Moulton is a true hero — a quality that they would like in a president; others, including many cable news reporters, focus on the negative comments and the deceptiveness of the video.

This scenario illustrates a situation in which a deepfake video is used by a PAC that is motivated to promote a candidate by valorizing his military record. It is used early in the campaign so there is time for reaction. One of the most salient ethical questions posed by this scenario is whether (and ultimately how) the synthesized nature of a video should be disclosed. The scenario is complicated on this matter because only one component of the video is synthesized. Another, perhaps more subtle issue has to do with the extent to which exaggerations of a candidate’s record are okay, particularly when the candidate has not consented. Exaggeration and inflation of a candidate’s qualifications has always been an issue in elections but deepfakes expand the possibilities for doing this.

Scenario #2

The race is neck and neck with only 3 days before the 2020 general election between Donald Trump and Elizabeth Warren. The winner may largely come down to who turns out to vote. Trump’s campaign advisors develop a strategy to get out the vote amongst his base: disaffected white voters. Campaign staff synthesize a deepfake video of Warren in a supposed close-door meeting with a few select members of the Black Congressional Caucus, and post it on Twitter and YouTube. In the cellphone-quality video she is heard saying disparaging and hateful things about white men in the US. In a matter of hours, Warren as well as other people falsely depicted in the video, publicly proclaim that the video is a fake. CNN and MSNBC quickly spread the word that the video is likely a fake. Nevertheless it spreads virally throughout social media, propelled and further amplified by troll and bot accounts. The video enrages Trump’s base, many of whom are unaware of the debunks or simply do not care, and spurs them to the polls for record turn-out on the right.

This scenario illustrates a situation in which a deepfake is used by campaign staff to hurt an opposing candidate by attributing to the opposing candidate extreme views that will incite the campaign’s base. Platforms play different roles here, as distributors and amplifiers not only of the lie, but also of its debunks. This scenario depicts the generic concern about the use of deepfakes, that is, that they will be used by one candidate (and/or their supporters) to distort and misrepresent a competitor candidate. In addition, misrepresenting a competitor can energize a candidate’s supporters. The scenario also points to the challenge of countering the effects of the deception especially when it is used late in a campaign. In the past candidates have mischaracterized their competitors, but deepfakes provide ostensible authenticity and credence to such mischaracterizations.

Scenario #3

The general election has come down to Pete Buttigieg and Donald Trump. While Trump has enjoyed strong evangelical support, Buttigieg has had difficulty garnering support from evangelicals. He was raised as a Catholic, is now a practicing Episcopalian, and in his campaign, has emphasized his Christian faith. In particular, he has repeatedly mentioned (in support of gay marriage) that his marriage (to a man) has made him a better Christian. Six weeks before the election, the internal polls of the Republican party begin to show that a small but not insignificant portion of evangelicals are moving towards Buttigieg. And then, ten days before the election, a deepfake video appears featuring several testimonials by men who say that they had sex with Mayor Pete while he has been married. A conservatively-aligned Political Action Committee (PAC) made the video and leaked it to a few evangelical groups via posts to their closed members-only Facebook groups using sock-puppet (i.e. fake) accounts. “Wow, look what I found — how can this guy say he’s christian?” reads one of the posts. Several religious leaders from Evangelical and other churches denounce Buttigieg to their congregations and followers.

As with others, this scenario illustrates how a deepfake might be used to harm a candidate by misrepresenting the candidate. However, in this case the responsible party is unknown, the strategy is to show the candidate to be a hypocrite, and instead of showing the candidate saying or doing something, the deepfake has others making statements as if they were real testimonials about a behavior for which there would be no other witnesses. The distribution of the deepfake is limited to a small number of people (members of a closed group). Hence, the scenario illustrates that deepfakes can have a significant impact even when the scale of distribution is relatively small. Indeed, with this mechanism of distribution, counteracting the effects is further challenged by their lack of public observability.

Scenario #4

Two weeks before the 2020 election, African-American voters throughout Iowa, Michigan, and Wisconsin receive emails spoofed to seemingly come from their local democratic party. The email addresses each recipient by name and tells them the location of their voting place. It also includes a personalized embedded video in which Reverend Al Sharpton addresses recipients by name and tells them that there have been attempts to undermine the election with false information about polling locations in their state. He encourages voters to use the information provided in the email when they go to vote. However, the video has been synthesized and the location information is bogus. On election day some voters are confused when they arrive to vote but are turned away because they’re at the incorrect location for their voter registration. The source of the deepfake is unknown until after the election when it is traced back to an alt-right group that would like to see Trump re-elected.

This scenario is unlike most of the others in that the deepfake is not used to attack or support a particular candidate but to undermine the democratic process by interfering with individuals’ autonomy in exercising their right to vote. It therefore indicates how deepfakes can be used to subvert the integrity of elections. Like the other scenarios, use of the deepfake involves deception, but not deception about candidates, rather deception about what an influential and recognized figure is claiming. The deepfake exploits the figure’s countenance to bolster trust in deceptive information and it raises complex questions about consent, property rights, and publicity rights around the use of an individual’s facsimile even if the individual is a public figure.

Scenario #5

Things are heating up and there’s a lot of competition from a wide group of contenders for the Democratic nomination. Bernie Sanders is campaigning heavily in Iowa and Nevada, two of the earliest primaries to be held, and races he knows he needs to win. A competitor to Sanders resolves to leak synthesized audio of Sanders saying disparaging things about Iowan farmers, calling them “hillbillies and rednecks” to a small group of mexican-american voters he met with in Nevada at a campaign stop. The audio clip gets uploaded onto Tumblr by an anonymous blogger who claims that he wants the world to know about “the real Bernie Sanders” but is worried about retaliation against himself and his family. The clip is picked up and played on several podcasts with followings on the left and right. The Sanders’ campaign is able to produce a complete recording of the purported event in Nevada which proves he never said those things. Major media outlets air debunkings of the clip, but it continues to circulate online and on some smaller podcasts.

In this scenario, an audiofake is anonymously produced and distributed to hurt a candidate by suggesting that the candidate is not who he purports to be, in the sense that he has said demeaning things about groups that he publicly supports. The case might be thought of as a simple case of using an audiofake to misrepresent a candidate; however, the scenario involves anonymity and raises a question about when anonymity is legitimate and how anonymous information should be treated. The case might also be seen in a somewhat positive light in that a real audio recording is used to counter the audiofake. However, it is difficult to say that the impact of the audiofake can be effectively counteracted by the real recording. How will listeners know which is real?

Scenario #6

It’s a tight race. Turn out in some key swing districts could make all the difference. Two days before the election, a foreign power unleashes a campaign to suppress certain voter demographics in battleground states. They send spoofed text messages with fake synthesized images to targeted individuals. The message says that if they vote on Tuesday, the attached image depicting the individual participating in a sex orgy with minors will be released publicly online and sent to their friends and family. These images are synthesized using a database of incriminating background scenes with a face swapped using photos of the targeted person scraped from their Facebook page or Instagram account. They are convincing enough to intimidate and coerce some people into not voting. A few individuals contact the police or their phone service provider, but the spoofed messages continue until election day. Since no one knows how many voters may have been affected, the incident undermines public perception of the legitimacy of the election.

In this scenario, we see a foreign actor interfering in an election campaign in a powerful way. The deepfake is not aimed at hurting or supporting a candidate, rather it is superficially aimed at voter suppression — superficial because the number of voters impacted would probably be quite small. Yet once the public becomes aware of this activity, the broader effect would be to cast doubt on the legitimacy of the election by suggesting voter suppression while making it difficult to understand its extent. In addition to eroding trust in election outcomes, this scenario shows how manipulated visual images can threaten individuals, and do so in private communication channels away from public observation. Like all the other scenarios, this one involves deception (because a synthesized video falsely represents behavior), but this one also illustrates how deepfakes can be used to coerce and intimidate.

Scenario #7

It’s the morning after the first presidential debate between Joe Biden and Donald Trump. Biden is basking in the media attention as the clear winner of the exchange when CNN suddenly airs an audio clip of the moderator of the debate which purports to be from an exchange the moderator had with a Biden campaign official before the debate. The clip was engineered by the Trump campaign to discredit Biden’s performance. In the audio the moderator is heard asking about whether “your candidate has any questions about the questions I sent yesterday?” Pundits interpret it as evidence of trying to fix the debate by sharing questions with Biden’s staff beforehand. The moderator, a respected journalist, firmly denies the account, but not before the #riggeddebates hashtag starts trending. Trump amplifies the idea that the debates are rigged to his Twitter following and refuses to participate in subsequent debates which are then cancelled.

This scenario illustrates how faked audio can be used simultaneously to hurt a competitor and undermine the integrity of (a component of) the election process, in this case a public debate. In this case, the attack was initiated by campaign staff suggesting corruption in the organization of the debate and participation in that corruption on the part of an opposing candidate. Historically, accusations of corruption have been used in campaigns and, to be credible, accusers have had to produce some sort of evidence to support their accusation. Deepfakes of the kind described in this scenario enable accusers to fabricate evidence that looks credible and and is widely distributed and amplified through social media. This expands the power of a false accusation both by making it seemingly real and by the quick and wide distribution that it has.

— — —

Disclosure: These scenarios were developed with support from the VisMedia Project.

--

--

Nick Diakopoulos

Northwestern University Professor of Communication. Computational journalism, algorithmic accountability, social computing — http://www.nickdiakopoulos.com/