This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

proofread

AI-generated arguments changed minds on controversial hot-button issues, according to study

AI's powers of political persuasion
Figure 1: Examples of AI and human-generated persuasive arguments.  Credit: Stanford University

Suddenly, the world is abuzz with chatter about chatbots. Artificially intelligent agents, like ChatGPT, have shown themselves to be remarkably adept at conversing in a very human-like fashion. Implications stretch from the classroom to Capitol Hill. ChatGPT, for instance, recently passed written exams at top business and law schools, among other feats both awe-inspiring and alarming.

Researchers at Stanford University's Polarization and Social Change Lab and the Institute for Human-Centered Artificial Intelligence (HAI) wanted to probe the boundaries of AI's political persuasiveness by testing its ability to sway real humans on some of the hottest social issues of the day—an assault weapon ban, the carbon tax, and paid parental leave, among others.

"AI fared quite well. Indeed, AI-generated persuasive appeals were as effective as ones written by humans in persuading human audiences on several political issues," said Hui "Max" Bai, a postdoctoral researcher in the Polarization and Social Change Lab and first author on a new paper about the experiment in pre-print.

Clever comparison

The research team, led by Robb Willer, a professor of sociology, psychology, and in the Stanford School of Humanities and Sciences and director of the Polarization and Social Change Lab, used GPT-3, the same large language model that fuels ChatGPT. They asked their model to craft persuasive messages on several controversial topics.

They then had thousands of real human beings read those persuasive texts. The readers were randomly assigned texts—sometimes they were written by AI, other times they were crafted by humans. In all cases, participants were asked to declare their positions on the issues before and after reading. The research team was then able to gauge how persuasive the messages had been on the readers and to assess which authors had been most persuasive and why.

Across all three comparisons conducted, the AI-generated messages were "consistently persuasive to human readers." Though the effect sizes were relatively small, falling within a range of a few points on a scale of zero to 100, such small moves extrapolated across a polarizing topic and a voting-population scale could prove significant.

AI's powers of political persuasion
Figure 2: Y-axes represent the difference between participants’ post-treatment and pre-treatment policy support. Higher scores indicate participants became more supportive of the policy. Credit: Stanford University

For instance, the authors report that AI-generated messages were at least as persuasive as human-generated messages across all topics, and on a smoking ban, gun control, carbon tax, an increased child tax credit, and a program, participants became "significantly more supportive" of the policies when reading AI-produced texts.

As an additional metric, the team asked participants to describe the qualities of the texts they had read. AI ranked consistently as more factual and logical, less angry, and less reliant upon storytelling as a persuasive technique.

High-stakes game

The researchers undertook their study of political persuasiveness not as a steppingstone to a new era of AI-infused but as a cautionary tale about the potential for things to go afoul. Chatbots, they say, have serious implications for democracy and for national security.

The authors worry about the potential for harm if used in a political context. Large language models, such as GPT-3, might be applied by ill-intentioned domestic and foreign actors through mis- or disinformation campaigns or to craft problematic content based on inaccurate or misleading information for as-yet-unforeseen political purposes.

"Clearly, AI has reached a level of sophistication that raises some high-stakes questions for policy- and lawmakers that demand their attention," Willer said. "AI has the potential to influence political discourse, and we should get out in front of these issues from the start."

Persuasive AI, Bai noted, could be used for mass-scale campaigns based on suspect information and used to lobby, generate online comments, write peer-to-peer text messages, or even produce letters to editors of influential print media.

"These concerning findings call for immediate consideration of regulations of the use of AI for political activities," Bai said.

More information: Artificial Intelligence Can Persuade Humans on Political Issues

Citation: AI-generated arguments changed minds on controversial hot-button issues, according to study (2023, March 2) retrieved 24 May 2024 from https://phys.org/news/2023-03-ai-generated-arguments-minds-controversial-hot-button.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Research finds Americans are more receptive to counter-partisan messages than previously thought

62 shares

Feedback to editors