News Release

Bots might prove harder to detect in 2020 elections

New study by USC researchers shows bots evolving to better mimic humans during elections

Peer-Reviewed Publication

University of Southern California

USC Information Sciences Institute (USC ISI) computer scientist, Emilio Ferrara, has new research indicating that bots or fake accounts enabled by artificial intelligence on social media have evolved and are now better able to copy human behaviors in order to avoid detection.

In the journal First Monday, research by Ferrara and colleagues Luca Luceri (Scuola Universitaria Professionale della Svizzera Italiana), Ashok Deb (USC ISI), Silvia Giordano (Scuola Universitaria Professionale della Svizzera Italiana), examine bot behavior during the US 2018 elections compared to bot behavior during the US 2016 elections.

The researchers studied almost 250,000 social media active users who discussed the US elections both in 2016 and 2018, and detected over 30,000 bots. They found that bots in 2016 were primarily focused on retweets and high volumes of tweets around the same message. However, as human social activity online has evolved, so have bots. In the 2018 election season, just as humans were less likely to retweet as much as they did in 2016, bots were less likely to share the same messages in high volume.

Bots, the researchers discovered, were more likely to employ a multi-bot approach as if to mimic authentic human engagement around an idea. Also, during the 2018 elections, as humans were much more likely to try to engage through replies, bots tried to establish voice and add to dialogue and engage through the use of polls, a strategy typical of reputable news agencies and pollsters, possibly aiming at lending legitimacy to these accounts.

In one example, a bot account posted an online Twitter poll asking if federal elections should require voters to show ID at the polls. It then asked Twitter users to vote and retweet.

Lead author, Emilio Ferrara, noted, "Our study further corroborates this idea that there is an arms race between bots and detection algorithms. As social media companies put more efforts to mitigate abuse and stifle automated accounts, bots evolve to mimic human strategies. Advancements in AI enable bots producing more human-like content. We need to devote more efforts to understand how bots evolve and how more sophisticated ones can be detected. With the upcoming 2020 US elections, the integrity of social media discourse is of paramount importance to allow a democratic process free of external influences."

###

To view the full study, please visit: https://firstmonday.org/ojs/index.php/fm/article/view/10213


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.