AI & Humanoids

AI offers improved civility for polarizing online conversations

AI offers improved civility for polarizing online conversations
Researchers have used AI to moderate polarizing online conversations
Researchers have used AI to moderate polarizing online conversations
View 2 Images
Researchers have used AI to moderate polarizing online conversations
1/2
Researchers have used AI to moderate polarizing online conversations
AI suggested rephrasing messages didn't alter the comment's content but provided options to the user to make a more polite statement
2/2
AI suggested rephrasing messages didn't alter the comment's content but provided options to the user to make a more polite statement

Researchers have used AI to create a way of improving the quality and civility of online discussions regarding polarizing topics by providing users with suggestions for rephrasing their comments before they post them. They say that, properly used, AI could be used to create a kinder, safer digital landscape.

Online conversations now play a central role in public discourse. But comment sections on social media platforms and digital news outlets are rife with discussions that have devolved into arguments, threats, and name-calling, particularly where the discussion concerns a divisive topic.

Now, researchers from Brigham Young University (BYU) and Duke University have developed AI that can moderate online chats, improving their quality and promoting civility.

They recruited 1,574 participants for their field experiment and asked them to engage in an online discussion about gun regulation in the US, a divisive issue that is often raised in the context of political debate. Each participant was matched with someone with an opposing view about gun policies.

Once matched, conversation pairs were randomly assigned to the treatment or control condition, and partners proceeded to have a conversation. In a treated conversation, one participant received three suggestions from GPT-3 for rephrasing their message before sending it. Participants could choose to send one of three AI-suggested alternatives, their original message, or edit any message.

AI suggested rephrasing messages didn't alter the comment's content but provided options to the user to make a more polite statement
AI suggested rephrasing messages didn't alter the comment's content but provided options to the user to make a more polite statement

On average, 12 messages were sent in each conversation, with a total of 2,742 AI-generated rephrasings suggested. Participants accepted AI-suggested rephrasings two-thirds of the time. Chat partners of individuals who implemented one or more AI rephrasing suggestions reported significantly higher conversation quality and were more willing to listen to the perspectives of their political opponents.

“We found the more often the rephrasings were used, the more likely participants were to feel like the conversation wasn’t divisive and that they felt heard and understood,” said David Wingate, one of the study’s co-authors.

The researchers say their findings suggest that this scalable solution could combat the toxic online culture that has pervaded the internet. They say it’d be easier to implement than, say, professional training sessions about online civility, which are limited in scope and availability as AI intervention could be broadly implemented across various digital channels.

Ultimately, say the researchers, this research shows that, properly used, AI can play an important role in creating a more positive online landscape, fostering discussions that are empathetic and respectful.

“My hope is that we’ll continue to have more BYU students build pro-social applications like this and that BYU can become a leader in demonstrating ethical ways of using machine learning,” Wingate said. “In a world that is dominated by information, we need students who can go out and wrangle the world’s information in positive and socially productive ways.”

The study was published in the journal PNAS.

Source: Brigham Young University

3 comments
3 comments
Daishi
This is such interesting research and while it could be deployed in a way that annoys people I do think there is opportunity to improve civil discourse in a lot of ways. I've seen forums with a "preview" option before "post" mostly for reasons of using formatting and markdown but having an AI validate and give pointers is also a decent enough use-case. I suspect it is a cultural thing but people online are generally terrible at debate and dealing with opinions they disagree with often resorting to name calling and ad hominem attacks instead of giving a reasoned based rebuttal. It's deeply unhealthy and I've had some more meaningful debates with AI on some topics than with people that actually succeeded in shifting my perspective.
paul314
Although this kind of tool could be used to increase civility, it's only ultimately useful if the participants in conversations incorporate some of those modifications into their views. Otherwise you might well end up with bad-faith trolling, where some people politely engage others in dialogs intended to stifle rather than advance serious discussions.
TM
This is very interesting, but... I can't help but also see this as extremely problematic. Of course, I completely agree with the fact that fighting against online toxicity/verbal violence is indispensable, and it's great to think about new ways to create more welcoming and positive spaces for meaningful conversations online. However I can't help but think that this type of solution could water down certain messaging, which I think is also a real risk.

I come from a culture where heated debates are the norm (I am French), which means that learning how to debate is something that is very strong in our education, starting at a young age. How to truly listen to what the other person is saying, analyse it and then answer back. I don't feel auto-correction is the solution, as the issue here is not vocabulary, style or even tone: what is truly at stake here is critical thinking, and I think that we need to further develop are indeed better tools and education materials/solutions to support that.