We think the Russian Trolls are still out there: attempts to identify them with ML

This is work done jointly by Jane Im, Eshwar Chandrasekharan, Libby Hemphill, David Jurgens, and Eric Gilbert. It is part of a new project at the Center for Social Media Responsibility at Michigan.

In the last two years, evidence has emerged that accounts controlled by the Russian Internet Research Agency (IRA) are operating on Twitter, likely with the goal of interfering with American democratic processes.

With the US 2018 election on our minds, we have built machine learning (ML) tools to label accounts still operating today as possible Russian trolls. Looking simultaneously at their behavior and language use, our machine learning model examines whether accounts operate in ways highly similar to the earlier group of trolls released by 538. While this is early work, we believe based on these findings that many new Russian troll accounts are likely active on Twitter, and currently working to shape American political discourse.

Summary of quantitative findings

Examples of ML-flagged accounts

1. @McCevich

@McCevich replying to Kevin Roose’s tweet about the interview.

Our model had, of course, no knowledge of this story, but processed the account because of the reply above. The model predicted @McCevich (his account, or one purporting to be him) as a troll with a probability of 100%. We believe this is face validity for the model’s decision boundary.

2. @jenni5262

Profile page of @jenni5262. The account was made in July 2017 and has tweeted over 18K times.
Two example @jenni5262 tweets. The account often tweets to news accounts about voting Republican.

3. @Rafael54356577

@Rafael54356577 targeting Kamala Harris over “illegals.”

The accounts aren’t simply bots.

The two accounts above do not pass the midpoint on Botometer.

While the ML-flagged accounts do exhibit some bot-like characteristics, it’s not a dominant behavior. This would seem essential to both defeat bot-detection software presumably already running on Twitter, as well as provide authenticity in conversations. One working hypothesis is that the Russian trolls might be software-assisted human workers.

Technical Details

Troll dataset. As part of the Mueller investigation, Twitter identified specific accounts controlled by the IRA; the news organization 538 obtained the list, and released data on these 2,848 Twitter accounts, all of whom were active during 2015–2017.

Distractor (“non-troll”) dataset. We assembled a set of approximately 170K “distractor accounts” to serve as counterexamples to the ones in the 538 dataset. These accounts are randomly selected Twitter accounts who had at sometime tweeted from the United States, and had tweeted at least 5 times over their lifetime on Twitter between 2012 and 2017. The list of these randomly selected accounts were provided by David Jurgens. In other words, they serve as “normal, everyday” accounts. Our model’s task is to distinguish the 2.8K known Russian troll accounts from this much larger set of 170K distractor accounts.

Method

  • Behavioral features: For example, the model looks at tweet frequency, the rate of retweeting, following count, etc.
  • Language features: In addition, the model examines both the distribution of languages present on an account’s timeline (e.g., English + Hungarian + Italian), as well specific term use from a dictionary of 5,000 highly important terms (e.g., “Trump”, “police”, “protest”, “MAGA”). These words were drawn from data, not imposed by the researchers a priori. In total, the model learned from over 17GB of quantitative and textual data.

Testing

Instead, you could ask: When the model makes a prediction of “likely Russian troll,” how often is it correct? This is known as “precision” in the field of machine learning. In our training data (the 538 dataset), the model is correct in this case 80% of the time.

Performance of the model using various metrics.

Predicting out-of-sample accounts
We expected the performance shown above to go down and it did when predicting “out-of-sample” — that is, predicting on new, unseen accounts in the real world. The unseen accounts were the 11,843 mentions of 11 high-profile journalists on Twitter. The model predicted 249 of the 11,843 to be trolls. 10 of the 11 journalists had at least one suspected Russian troll trying to get their attention in their recent mentions. We believe after human review that 50–70% of these model-flagged accounts are highly likely to be Russian trolls (i.e., 13–18 per journalist). This is a large range, reflecting the early and fluid nature of this project.

Conclusion

We would love feedback on this work. As journalists often use tweets in their practice, we would be happy to run this model one-off on their mentions in the run-up to the election. Please feel free to contact Eric Gilbert by email (available on linked site) to answer any further questions about the work.

Eric Gilbert is the John Derby Evans Associate Professor in the School of Information, and Professor of EECS, at the University of Michigan.

David Jurgens is an Assistant Professor in the School of Information and EECS at the University of Michigan.

Libby Hemphill is an Associate Professor in the School of Information, a Research Associate Professor at ISR, and the Director of the Resource Center for Minority Data at ICPSR, all at Michigan.

Eshwar Chandrasekharan is a PhD student in the School of Interactive Computing at Georgia Tech. He is advised by Eric Gilbert.

Jane Im is a PhD student in the School of Information at the University of Michigan. She is co-advised by David Jurgens and Eric Gilbert.

I’m a PhD student at the University of Michigan School of Information and CSE. I study and build social computing systems. :) http://imjane.net