A new study has predicted that AI activity by ‘bad actors’ determined to cause online harm through the spread of disinformation will be a daily occurrence by the middle of 2024. The findings are concerning given that more than 50 countries, including the US, will hold national elections this year, the outcomes of which will have a global impact.
Even before the release of the newest iterations of Generative Pretrained Transformer (GPT) systems, AI experts were forecasting that by 2026, 90% of online content will be generated by computers without human intervention, giving rise to the spread of mis- and disinformation.
There’s an assumption that big social media platforms with the largest number of users warrant regulation to rein in risk. That assumption is correct, to a degree, and makes them a target of legislation such as the EU’s Digital Services Act and AI Act. However, there are other, smaller ‘bad actors’ – people, groups and countries who purposely engage in behavior that causes harm to others – who misuse AI.
A new study led by researchers at George Washington University (GW) is the first quantitative scientific analysis to examine how bad actors might misuse AI and GPT systems to generate harm globally across social media platforms and what can be done about it.
“Everybody is talking about the dangers of AI, but until our study there was no science behind it,” said Neil Johnson, lead author of the study. “You cannot win a battle without a deep understanding of the battlefield.”
The researchers started by mapping the dynamic network of interlinking social media communities that make up the landscape of the global online population. Users – sometimes a few, sometimes a few million – join these communities because of shared interests, which can include harms. The researchers focused on extreme ‘anti-X’ communities where each community is one in which two or more of its 20 most recent posts include defined hate speech and/or extreme nationalism and/or racism. These anti-X communities included those that were, for example, anti-US, anti-women, anti-abortion, or anti-semitic. Links between these communities form over time to create clusters of communities within and across different social media platforms.
“Any community A may create a link (i.e. a hyperlink) to any community B if B’s content is of interest to A’s members,” said the researchers. “A may agree or disagree with B. This link directs A’s member's attention to B, and A’s members can then add comments on B without B’s members knowing about the link – hence, community B’s members have exposure to, and potential influence from, community A's members.”
Using a mathematical model, the researchers determined what bad-actor-AI activity is likely to occur and why. Specifically, they found that the most basic GPT system, such as GPT-2, is all that is needed and is also more likely to be attractive to bad actors than more sophisticated versions, such as GPT-3 or -4. This is because GPT-2 can easily replicate the human style and content already seen in extreme online communities, and bad actors can use a basic tool like GPT-2 to produce more inflammatory output by subtly changing the form of an online query without changing its meaning. By contrast, GPT-3 and -4 contain a filter that overrides answers to potentially contentious prompts, preventing such output.
The online ‘battlefield’ where bad-actor-AI activity will likely thrive, say the researchers, is the bad-actor communities plus the communities they directly link into, that is, vulnerable mainstream communities. Adding these communities together amounts to an online ecosystem of more than one billion individuals, allowing bad-actor-AI to thrive globally. The researchers illustrate their point by referencing the non-AI-generated hate and extremism related to COVID-19 and, more recently, the Russia-Ukraine and Israel-Hamas wars.
They predict that, by mid-2024, bad-actor-AI activity will become a daily occurrence. To determine this, they used proxy data from two historical, technologically similar incidents involving the manipulation of online electronic systems: the 2008 automated algorithm attacks on US financial markets and the 2013 Chinese cyber-attacks on US infrastructure. Analyzing these data sets, they extrapolated the frequency of attacks in these two events in the context of the current technological progress of AI.
2024 is being touted as the ‘biggest election year in history,’ with more than 50 countries, including the US, due to hold national elections this year. From Russia, Taiwan, the UK, and India to El Salvador and South Africa, the outcomes of some elections will have a global impact and huge implications for human rights, economies, international relations, and world peace. So, say the researchers, the threat of bad actors using AI to disseminate and amplify disinformation during these elections is real.
They recommend that social media companies use tactics to contain disinformation rather than removing every piece of bad-actor-generated content.
Given AI's ever-changing landscape, the researchers did apply a caveat to their study’s findings. Nonetheless, the study highlights some of the significant challenges posed by bad actors with access to AI.
“Since nobody can predict exactly what will happen with future bad-actor-AI given the rapid pace of technology and changing online landscape, the predictions in this article are, strictly speaking, speculative,” the researchers said. “But they are each quantitative and testable – and also generalizable – and hence provide a concrete starting point for strengthening bad-actor-AI policy discussions.”
The study was published in the journal PNAS Nexus.
Source: GW