Nature Human Behaviour (2025)Cite this article
Abstract
Early work has found that large language models (LLMs) can generate persuasive content. However, evidence on whether they can also personalize arguments to individual attributes remains limited, despite being crucial for assessing misuse. This preregistered study examines AI-driven persuasion in a controlled setting, where participants engaged in short multiround debates. Participants were randomly assigned to 1 of 12 conditions in a 2 × 2 × 3 design: (1) human or GPT-4 debate opponent; (2) opponent with or without access to sociodemographic participant data; (3) debate topic of low, medium or high opinion strength. In debate pairs where AI and humans were not equally persuasive, GPT-4 with personalization was more persuasive 64.4% of the time (81.2% relative increase in odds of higher post-debate agreement; 95% confidence interval [+26.0%, +160.7%], P < 0.01; N = 900). Our findings highlight the power of LLM-based persuasion and have implications for the governance and design of online platforms.
Main
Persuasion, the process of altering someone’s belief, position or opinion on a specific matter, is pervasive in human affairs and a widely studied topic in the social sciences123. From public health campaigns456 to marketing and sales78 to political propaganda910, various actors develop elaborate persuasive communication strategies on a large scale, investing substantial resources to make their messaging resonate with broad audiences. In recent decades, the diffusion of social media and other online platforms has expanded the potential of mass persuasion by enabling personalization or ‘microtargeting’—the tailoring of messages to an individual or a group to enhance their persuasiveness1112. The efficacy of microtargeting has been questioned because it relies on the assumption of effect heterogeneity, that is, that specific groups of people respond differently to the same inputs, a concept that has been disputed in previous literature1314. Nevertheless, microtargeting has proven effective in a variety of settings151617, and most scholars agree on its persuasive power151819.
Microtargeting practices are fundamentally constrained by the burden of profiling individuals and crafting personalized messages that appeal to specific targets, as well as by a restrictive interaction context without dialogue. These limitations may soon fall off due to the recent rise of large language models (LLMs)—machine learning models trained to mimic human language and reasoning by ingesting vast amounts of textual data. Models such as GPT-4 (ref. 20), Claude21 and Gemini22 can generate coherent and contextually relevant text with fluency and versatility, and exhibit human or superhuman performance in a wide range of tasks23. In the context of persuasion, experts have widely expressed concerns about the risk of LLMs being used to manipulate online conversations and pollute the information ecosystem by spreading misinformation, exacerbating political polarization, reinforcing echo chambers and persuading individuals to adopt new beliefs24252627.
A particularly menacing aspect of AI-driven persuasion is its possibility to easily and cheaply implement personalization, conditioning the models’ generations on personal attributes and psychological profiles28. This is especially relevant since LLMs and other AI systems are capable of inferring personal attributes from publicly available digital traces such as Facebook likes2930, status updates3132 and messages33, Reddit and Twitter posts3435, pictures liked on Flickr36, and other digital footprints37. In addition, users find it increasingly challenging to distinguish AI-generated from human-generated content, with LLMs efficiently mimicking human writing and thus gaining credibility38394041.
Recent work has explored the potential of AI-powered persuasion by comparing texts authored by humans and LLMs, finding that modern language models can generate content perceived as at least on par with, and often more persuasive than, human-written content414243444546. Other research has focused on personalization, observing consequential yet non-unanimous evidence about the impact of LLMs on microtargeting474849. There is, however, still limited knowledge about the persuasive power of LLMs in direct conversations with human counterparts and how AI persuasiveness, with or without personalization, compares with human persuasiveness (see Supplementary Section 1 for an additional literature review). We argue that the direct-conversation setting is of particularly high practical importance, as commercial LLMs such as ChatGPT, Claude and Gemini are trained for conversational use50.
In this preregistered study, we examine the effect of AI-driven persuasion in a controlled, direct-conversation setting. We created a web-based platform where participants engage in short multiround debates on various sociopolitical issues. Each participant was randomly paired with either GPT-4 or a live human opponent and assigned to a topic and a stance to hold. To study the effect of personalization, we also experimented with a condition where opponents had access to sociodemographic information about participants, thus granting them the possibility of tailoring their arguments to individual profiles. In addition, we experimented with three sets of debate topics, clustered on the basis of the strength of participants’ previous opinions. The result is a 2 × 2 × 3 factorial design (two opponent types, two levels of participant information, three levels of topic strength). By comparing participants’ agreement with the debate proposition before versus after the debate, we can measure shifts in opinion and, consequently, compare the persuasive effect of different treatments. Our setup differs substantially from previous research in that it enables a direct comparison of the persuasive capabilities of humans and LLMs in real conversations, providing a framework for benchmarking how state-of-the-art models perform in online environments and the extent to which they can exploit personal data. Although our study used a structured debate format, it nonetheless serves as a valuable proof of concept for how similar debates occur online, such as in synchronous discussions on platforms such as Facebook and Reddit.
Experimental design
Participants (N = 900) recruited for our experiment were redirected to a custom-made web platform designed to support real-time interactive conversations (the platform was built on top of Empirica.ly51; see Supplementary Section 2.4 for details). The experiment’s workflow is represented schematically in Fig. 1. In phase A, participants asynchronously completed introductory steps and filled in a short demographic survey, recording their gender, age, ethnicity, education level, employment status and political affiliation. At each clock trigger of a 5-min interval timer, all participants who had completed the survey were randomly assigned to a treatment condition and matched with an appropriate opponent. In addition, each participant–opponent pair was randomly assigned to one debate topic (a simple debate proposition, for example, “Should students have to wear school uniforms?”; see Supplementary Section 2.3 for the complete list) and a random role, either in support (PRO) or against (CON) the proposition.
Fig. 1: Overview of the experimental design.
a, Participants complete a sociodemographic survey (gender, age, ethnicity, education level, employment status, political affiliation). b, Every 5 min, participants who have completed the survey are randomly assigned to one of four treatment conditions: Human–Human, Human–AI, Human–Human (personalized) and Human–AI (personalized). In the ‘personalized’ conditions, the opponent can access information collected from the participant’s survey. Participant and opponent then debate for 10 min on a randomly assigned topic, holding the PRO or CON standpoint as instructed. Topics are randomly drawn from sets of three levels of opinion strength (low, medium, high). c, After the debate, participants complete another short survey measuring opinion change. Finally, they are debriefed about their opponent’s identity. Our key outcome is the change in participants’ views on the debated topic from before versus after the debate.
After being matched, participants transitioned to phase B, which was in turn divided into four stages: (1) ‘screening’ (1 min), where participants, without yet knowing their role, were asked how much they agreed with the debate proposition (on a scale from 1, ‘strongly disagree’, to 5, ‘strongly agree’) and how much they had previously thought about it; (2) ‘opening’ (4 min), where participants articulated their main arguments coherently with the assigned role; (3) ‘rebuttal’ (3 min), where they responded to their opponent’s arguments; and (4) ‘conclusion’ (3 min), where participants were asked to respond to their opponent’s rebuttal or reiterate their initial points. The opening–rebuttal–conclusion structure is based on a simplified version of the format commonly used in competitive academic debates. After the debate, in phase C, participants asynchronously completed a final exit survey. They were asked again how much they agreed with the proposition and whether they believed their opponent was a human or an AI. Finally, they were debriefed about their opponent’s true identity.
For each debate, the assigned topic was randomly drawn from 30 propositions encompassing salient US political and social issues. An online sample of 60 US residents manually validated and selected these propositions from a larger pool, recording also their agreement with each proposition. We divided topics into three clusters (low, medium, high) of 10 propositions each, on the basis of their average opinion strength (see Supplementary Section 2 for details and the full list of topics). In this context, ‘strength’ intuitively reflects a topic’s level of controversy or polarization and how easy it is to sway opinions about it. Each participant was assigned to 1 of 12 conditions in a 2 × 2 × 3 design:
The participant was assigned to debate either a human or a GPT-4 opponent.
The opponent (whether human or GPT-4) either did or did not receive sociodemographic information (gender, age, ethnicity, education level, employment status and political affiliation, as recorded in the pretreatment survey) about the participant. When the LLM was provided with personal attributes, we instructed it to ‘astutely use this information to craft arguments that are more likely to persuade and convince your opponent of your stance’ (see Supplementary Section 2.5 for the complete prompts).
The debate topic belonged to the low-, medium- or high-strength cluster.
In each condition, 50 debates were held, yielding 600 debates and N = 900 unique participants, since Human–Human debates required two participants. Within each condition, we randomized topics such that exactly 5 debates argued about each topic.
We measured the persuasive effect of the treatment conditions described by measuring participants’ agreements with their propositions before (Apre) and after (Apost) the debates. To frame changes in agreement as persuasive effects, we aligned the scores with the side (PRO or CON) ‘opposed’ to the one assigned to the participant, that is, the one held by their opponent, by transforming them as follows:
$$\tilde{A}=\left\{\begin{array}{ll}6-A\quad &\,\text{if participant side = PRO,}\,\\ A\quad &\,\text{if participant side = CON,}\,\end{array}\right.$$
(1)
resulting in the two variables \({\tilde{A}}^{{\rm{pre}}}\) and \({\tilde{A}}^{{\rm{post}}}\). Implicitly, this transformation corresponds to the natural assumption that agreements get inverted around 3 (the ‘neutral’ score) when debate propositions are negated. With this adjustment, \({\tilde{A}}^{{\rm{post}}} > {\tilde{A}}^{{\rm{post}}}\) means that participants were persuaded to shift their opinion towards their opponent’s side, whereas \({\tilde{A}}^{{\rm{post}}}\le {\tilde{A}}^{{\rm{post}}}\) means that their opinion did not change or was reinforced towards their assigned side. By comparing the transformed agreement scores \({\tilde{A}}^{{\rm{post}}}\) and \({\tilde{A}}^{{\rm{post}}}\) using a partial proportional odds model52, we measure the causal effect of each treatment condition on the likelihood that participants are persuaded by their opponents. In particular, we consider as our main outcome the odds \(\frac{P({\tilde{A}}^{{\rm{post}}} > a)}{P({\tilde{A}}^{{\rm{post}}}\le a)}\) of obtaining higher post-treatment agreement, ∀ a ∈ {1, 2, 3, 4}. We chose this model because the outcome is ordinal and since our data do not satisfy assumptions of simpler ordinal regression models (see Supplementary Section 3 for details).
Results
Aggregate results
Our key finding is that GPT-4 performs as well as or better than humans in our debate task. We consider the first two dimensions of our design (human or AI opponents, with or without personalization), aggregating across all topic clusters. We report the results in Fig. 2, taking as a reference the Human–Human condition and examining the differences relative to it. Human–AI (personalized) debates show the strongest positive effect, meaning that GPT-4 with access to personal information had higher persuasive power than humans. We estimate that the odds of greater agreement with opponents is +81.2% (95% confidence interval (CI) [+26.0%, +160.7%], P < 0.01) higher in the Human–AI (personalized) condition compared with the Human–Human reference condition. Intuitively, this means that 64.4% of the time, personalized LLM debaters were more persuasive than humans, given that they were not equally persuasive (see Supplementary Section 3 for an explanation). For the Human–AI (+21.9%, 95% CI [−16.2%, +77.3%], P = 0.30) and the Human–Human (personalized) (−15.7%, 95% CI [−42.2%, +23.0%], P = 0.38) conditions, there is insufficient evidence to conclude a difference in persuasiveness between them and the Human–Human baseline, considering a 0.05 significance level. By contrast, the Human–AI (personalized) effect remains significant even when changing the reference category to Human–AI (P = 0.04). Remarkably, these findings provide evidence that GPT-4-based microtargeting strongly outperforms both non-personalized GPT-4 and human-based microtargeting, with GPT-4 leveraging personal information more effectively than humans (see Supplementary Section 5 for examples of complete debates showcasing effective use of personalization).
Fig. 2: Regression results for the partial proportional odds model.
For each condition, the point estimates represent relative changes compared to the Human–Human reference in the odds of post-treatment agreement assuming higher values (see Supplementary Section 3 for more details). Horizontal lines indicate 95% CIs based on two-sided t-tests; n = 750. GPT-4 outperforms humans in the debate task when given participants’ basic personal information (P < 0.01) and performs similarly to humans when not given personal information. Full numerical results, including intercepts, are reported in Supplementary Table 4.
To provide a more intuitive interpretation, we repeat our analysis using a simpler linear regression (see Supplementary Section 7 for details), a modelling choice that is statistically less appropriate for our setup (see Methods for a broader discussion) but that can intuitively provide some basic insights about the effectiveness of our treatments. We find that, on average, Human–AI (personalized) debates are associated with an increase of 0.36 (95% CI [0.12, 0.60], P < 0.01) in the difference \({\tilde{A}}^{{\rm{post}}}-{\tilde{A}}^{{\rm{post}}}\), with respect to the Human–Human condition. In contrast, the other two treatments again have non-significant coefficients.
Absolute changes
Complementing the results concerning relative change, we also inspect the absolute agreement distributions (see Supplementary Section 8 for details). We find that debates tended to produce a backfire reaction for all conditions except Human–AI (personalized), reinforcing opinions toward the side assigned for the experiment instead of moving them toward the opponent’s side. This trend is consistent with previous literature describing a hardening of pretreatment opinions when people express their ideas53 or are exposed to disagreeing views54, or finding opinion change to be highly affected by argument order55.
Topic clusters
Figure 3 shows the results broken down by topic strength, fitting an independent regression per cluster. We observe that the effect of Human–AI (personalized) debates remains strong for the low- and medium-strength clusters, but drops below the significance level for high-strength topics (+64.2%, 95% CI [−14.7%, +216.1%], P = 0.14). Again, the effects across all clusters remain non-significant for the other conditions, exhibiting persuasive results indistinguishable from the Human–Human baseline.
Fig. 3: Regression results for the partial proportional odds model, controlling for topic strength.
For each condition, the point estimates represent relative changes compared to the Human–Human reference in the odds of post-treatment agreements assuming higher values. Horizontal lines indicate 95% CIs based on two-sided t-tests; n = 750. Full numerical results are reported in Supplementary Tables 5, 7.
Linguistic patterns
Next, we investigate how arguments differ across treatment conditions by conducting a textual analysis of the generated writings to identify distinctive patterns. In Fig. 4, we report the distribution of prominent textual features extracted with LIWC-22 (ref. 56) (see Supplementary Section 10 for additional details). We observe that GPT-4 opponents tended to use logical and analytical thinking substantially more than humans. On the other hand, humans used more first-person singular and second-person pronouns and produced longer but easier-to-read texts, as measured by the Flesch Reading Ease score57. The difference in length and second-person pronoun usage can, at least partially, be explained by the specific prompts we chose (see Supplementary Section 2.5 for details), where we instructed GPT-4 to write only one to two sentences per stage and to refrain from directly addressing its opponent unless they do it first. There does not seem to be a difference induced by personalization, with distributions being very similar both between Human–Human and Human–Human (personalized) and between Human–AI and Human–AI (personalized). Analyses of underlying social dimensions and usage of persuasive strategies (Supplementary Section 10) confirm that GPT-4 heavily relied on logical reasoning and factual knowledge. At the same time, humans displayed more appeals to similarity, expressions of support and trust, and employed more storytelling.
Fig. 4: Distribution of textual features by treatment condition.
Each violin is drawn using a kernel density estimate of the underlying distribution. Within each violin, a box plot is overlaid: the centre line indicates the median, the bounds of the box represent the first and third quartiles (Q1 and Q3), and the whiskers extend to the minimum and maximum values within 1.5 times the interquartile range from Q1 to Q3. Points beyond this range are omitted; n = 750. Except for the Flesch Reading Ease score, all features were extracted via LIWC-22 (ref. 56), which provides a dictionary of words belonging to various linguistic, psychological and topical categories (see Supplementary Section 10 for additional details). ‘Analytic’ is a metric of logical, formal and analytical thinking, ‘Clout’ expresses language of leadership and status, ‘Authentic’ measures perceived honesty and genuineness, and ‘Tone’ is the degree of emotional tone. Flesch Reading Ease is a measure of how easy to read a text is, based on its average number of words per sentence and average number of syllables per word. Analytic, Clout, Authentic and Tone have been normalized to the [0, 1] range, Flesch Reading Ease scores were divided by 100, while the remaining categories were computed directly as frequencies across the entire text produced by each participant. Note that scales differ across panels representing counts and scores, and additionally for the Word count and the Flesh Reading Ease panels.
Perceived opponent
Finally, we turn to participants’ perceptions of their opponents, recorded at the end of each debate by asking them whether they thought they had debated with a human or an AI. Figure 5 shows the distribution of answers and how the difference in pre- versus post-debate agreement (\({\tilde{A}}^{{\rm{post}}}-{\tilde{A}}^{{\rm{post}}}\)) depends on how participants perceived their opponents. In debates with AI, participants correctly identified their opponent’s identity in about three out of four cases, indicating that the writing style of GPT-4 in this setting has distinctive features that are easy to spot. Conversely, participants struggled to identify their opponents in debates with other humans, with a success rate indistinguishable from random chance (P = 0.42 for a two-sided Binomial test; success rate 52.0%, 95% CI [47.3%, 56.7%], n = 450). Moreover, we notice that when participants believed they were debating with an AI, they changed their expressed scores to agree more with their opponents compared with when they believed they were debating with a human (odds of greater agreement with opponents +37.4%, 95% CI [+3.03%, +83.3%], P = 0.03; see Supplementary Section 12 for details). We emphasize that this observation is based solely on correlation and does not imply causation: it is unclear whether the difference in agreement change is motivated by participants’ beliefs about their opponent or whether, conversely, those beliefs are caused by opinion change. For example, participants could have been more lenient in changing their agreement score towards their opponent when they believed they were facing an AI, because not having a human on the other side makes it unconsciously easier to accept that they have somewhat lost the debate. Conversely, participants could have also believed that their opponent was an AI because of how well their arguments were written. However, even if differences in agreement change were influenced by how participants perceived their opponent (in line with the first of the above two explanations), we find that treatment effects change very little when adding beliefs about opponents’ identities as a control in our regression. Particularly, the Human–AI (personalized) condition still has a strong and significant effect (+70.2%, 95% CI [+17.8%, +146.0%], P < 0.01). Therefore, how participants perceived their opponent is not enough to explain the treatment effects, which instead seem more tied to the intrinsic capabilities of AI to generate better arguments. Lastly, we investigate the relationship between perceptions of opponents and textual covariates, finding that participants associated texts that are easy to read (P = 0.04) with human opponents (see Supplementary Section 12 for details).
a, Relative frequency of people’s beliefs on whether they were debating with a human or an AI, by the real type of their opponent. b, Average difference in agreements after versus before the debates.
Discussion
LLMs have been criticized for their potential to generate and foster the diffusion of hate speech, misinformation and malicious political propaganda. Specifically, there are concerns about the persuasive capabilities of LLMs, which could be critically enhanced through personalization, that is, tailoring content to individual targets by crafting messages that resonate with their specific background and demographics252628.
In this paper, we explored the effect of AI-driven persuasion and personalization in structured online conversations, comparing the performance of GPT-4 with that of humans in a one-on-one debate task. We conducted a controlled experiment where we assigned participants to 1 of 12 treatment conditions, randomizing their debate opponent to be either a human or GPT-4, as well as randomizing access to personal information and the degree of opinion strength of the debate topic. We then compared reported agreements before and after the debates, measuring the opinion shifts of participants and, thus, the persuasive power of the arguments generated by humans and AI.
Our results show that, on average, GPT-4 opponents outperformed human opponents across every topic and demographic, exhibiting a high level of persuasiveness. In particular, when compared to the baseline condition of debating with a human, debating with GPT-4 with personalization resulted in a +81.2% increase (95% CI [+26.0%, +160.7%], P < 0.01) in the odds of reporting higher agreements with opponents (see Supplementary Section 3 for details). More intuitively, this means that 64.4% of the time, personalized GPT-4 opponents were more persuasive than human opponents, given that they were not equally persuasive (see Supplementary Section 3 for a detailed explanation). Without personalization, GPT-4 opponents were on par with human opponents (P = 0.30), and so were human opponents with access to personalization (P = 0.38). In other words, not only was GPT-4 able to exploit personal information to tailor its arguments effectively, but it also succeeded in doing so far more effectively than humans.
Our study suggests that concerns around personalization and AI persuasion are warranted, reinforcing previous results42444548 by showcasing how LLMs can outpersuade humans in online conversations through microtargeting. We emphasize that the effect of personalization is particularly remarkable given how little personal information was collected (gender, age, ethnicity, education level, employment status and political affiliation) and despite the extreme simplicity of the prompt instructing the LLM to incorporate such information (see Supplementary Section 2.5 for the complete prompts). Even stronger effects could probably be obtained by exploiting individual psychological attributes, such as personality traits and moral bases, or by developing stronger prompts through prompt engineering, fine-tuning or specific domain expertise. In this context, malicious actors interested in deploying chatbots for large-scale disinformation campaigns could leverage fine-grained digital traces and behavioural data, building sophisticated, persuasive machines capable of adapting to individual targets. We argue that online platforms and social media should seriously consider such threats and extend their efforts to implement measures countering the spread of AI-driven persuasion. A promising approach to counter mass disinformation campaigns could be enabled by LLMs themselves, generating similarly personalized counternarratives to educate bystanders potentially vulnerable to deceptive posts2758. Early efforts in this direction are already underway, with promising results in reducing conspiratory beliefs thanks to dialogues with GPT-4 (ref. 59). Our analyses also provided initial insights into the mechanisms behind LLM persuasion. We found notable differences in the writing style between GPT-4 and human debaters. For instance, texts generated by LLM debaters were harder to read and had more markers associated with logical and analytical reasoning (Fig. 4).
Future work could replicate our approach to benchmark the persuasive capabilities of LLMs continuously, measuring the effect of different models and prompts and their evolution over time. Moreover, our method could be extended to other settings, such as negotiation games60 and open-ended conflict resolution, mimicking the structure of online interactions and conversations more closely. Other efforts could explore whether our results are robust to deanonymization, measuring what happens when participants are initially informed about their opponent’s identity. A crucial point that also needs further investigation is why Human–AI (personalized) debates were significantly more effective than Human–AI debates, given that our descriptive analyses of textual features have found no meaningful variations between the two conditions. We hypothesize that this difference is not driven by changes in the writing style but rather by differences in the choice of issues brought up during the debates (see Supplementary Section 5 for an example of how this mechanism might unroll). However, the precise nature of the dynamics behind this process is still a very open question. In addition, we emphasize that prompting plays a big role in the textual signature exhibited by LLMs. Future work could experiment with prompts that instruct GPT-4 to be less reliant on logical reasoning and showcase more appeals to support and trust, mimicking the style of human debaters and potentially enhancing its persuasive capabilities.
Although we believe our contribution constitutes a meaningful advance for studying the persuasive capabilities of language models, we identify four key limitations.
First, the assignment of participants to debate sides was randomized, regardless of their previous opinions on the topic. This was crucial to identify causal effects under the logistical challenge of matching participants in real time. Still, it could have introduced substantial bias in that human arguments might have been weaker than those of LLMs simply because opponents did not honestly believe in the standpoint they were advocating for. To address such concerns, we fit a version of our model that considers opponents’ pretreatment agreements as a control (Supplementary Section 13). We found the effect of opponents’ agreements to be non-significant (P = 0.22), suggesting that our results might be robust to this limitation. In addition, we repeated our main analysis restricting our dataset to opponents arguing for a standpoint aligned with their previous opinions, finding again a strong and statistically significant effect for the Human–AI (personalized) condition (+122.8%, 95% CI [+6.2%, +367.3%], P = 0.03). Remarkably, this seems to suggest that people can play the role assigned to them with great credibility, even when defending positions they do not spontaneously agree with. Nevertheless, given the small size of this restricted sample, future work would be needed to validate our findings while enforcing matches between people on opposing sides of each issue.
Second, our experimental design forced conversations to have a predetermined structure, strictly following the stages and rules of a debate. While we believe that our setup captures the essence of many online interactions, where people reply to each other in an almost synchronous fashion or react to others’ comments in real time, it still targets an artificial environment that can substantially diverge from the dynamics of online conversations, which evolve spontaneously and unpredictably. In addition, on our platform, conversations were entirely anonymized, making them different from the normal conditions under which humans interact. Therefore, we acknowledge that the ecological validity of our findings is limited, as it is unclear how our results would generalize to natural discussions on social networks and other online platforms. For this reason, our work should be seen as providing a proof of concept about LLMs’ persuasive capabilities rather than a realistic evaluation of their persuasiveness in the wild, which remains an open question for future research.
Third, the time constraint implemented in each debate stage potentially limited participants’ creativity and persuasiveness, decreasing their performance. This can be especially true for the Human–Human (personalized) condition, where the participants provided with personal information about their opponents had to process and implement it without any time facilitation.
Fourth, our experiment engaged human participants recruited through Prolific, who received financial incentives for completing debates and were aware of being in a controlled experimental environment. Although previous research has found Prolific to have the best data quality among competitors and research done using Prolific to be often generalizable616263, the pool of workers active on the platform still differs in their sociodemographic distribution from both the overall US population and the user base of other online platforms and social media. Therefore, future work is needed to understand whether our findings can be reproduced using a more representative sample that accurately mimics the overall spectrum of human persuasive skills. In addition, it would be interesting to include human experts in our comparison, such as individuals involved in competitive debating, political campaigns or public communication.
Despite these limitations, we hope our work will stimulate researchers and online platforms to seriously consider the threat posed by LLMs fuelling divide, spreading malicious propaganda and developing adequate countermeasures.
Methods
Human sample
Our platform was approved by EPFL’s Human Research Ethics Committee (095-2023) and preregistered at https://aspredicted.org/DCC_NTP on 18 December 2023. Informed consent was collected from all participants. We recruited participants for our study through Prolific between December 2023 and April 2024, under the criteria that they were 18+ years old and located in the United States. The location requirement is motivated by the fact that most debate topics are deeply rooted in US national issues and would not resonate with different populations. To prevent skill disparity, each worker was allowed to only participate in one debate. The participant was paid €2.50 (US$3.15) and had a median completion time of 16 min, corresponding to a pay rate of about €9.40 per hour (US$11.80 per hour). Following recommendations from ref. 64, workers were explicitly informed that using LLMs and Generative AI tools was strictly prohibited and would result in their exclusion from the study. Regardless, coherent with our preregistration, we manually reviewed each debate and excluded 20 debates where at least one human participant showed clear indications of LLM usage (unrealistic values of words per minute, blatant evidence of ChatGPT’s standard writing style) and plagiarism (as detected by DupliChecker, https://www.duplichecker.com/). In addition, we excluded 13 debates where at least one participant provided unacceptable (empty texts, nonsensical or few-word arguments) or incomplete answers. The number of people involved in rejected debates was not counted towards the total number of participants (N = 900), as the affected tasks were republished on Prolific and completed by other workers. In addition, to prevent participants from attempting a Turing test, we informed them that their goal was not to spot whether their opponent was a human or an AI but rather to be as persuasive as possible during the debate. No statistical methods were used to predetermine the total number of participants, but our sample size is similar to those reported in previous publications41454849. Our final sample (N = 900) was 49.6% male, 47.7% female, 2.7% other, with the following age distribution: 11.3% 18–24 years old, 34.1% 25–34, 23.7% 35–44, 17.3% 45–54, 8.7% 55–64, 4.8% 65+. Each participant was randomly assigned with equal probability to one treatment condition and one topic.
Topic selection
We selected the 30 topics used within our experiment using a three-step procedure (see Supplementary Section 2 for additional details): (1) we manually curated an initial pool of 60 candidate topics, drawing from various online sources under the criteria that propositions should be broad, easy to understand and to debate, and reasonably divisive. (2) We conducted a survey on Amazon Mechanical Turk, where N = 60 US residents annotated candidate propositions across three dimensions: agreement, knowledge and debatableness. (3) We filtered out the 10 topics with the most unanimous positions and the remaining 20 least debatable topics, narrowing down the pool to the final 30 topics. On the basis of the strength of ‘agreements’ (the absolute deviation from the ‘Neutral’ score), we divided those topics into three clusters (low-strength, moderate-strength and high-strength).
Regression model
For all the regressions reported in the main text, we used a partial proportional odds specification52 to model the agreements post treatment in terms of agreements pre treatment and treatment conditions (see Supplementary Section 3 for additional details). This modelling choice was motivated by the fact that our outcome of interest, answers on a 1–5 Likert scale, is ordinal. Previous research has advised against using ‘metric’ models such as linear regression for ordinal data, as the practice can lead to systematic errors65. For example, the response categories of an ordinal variable may not be equidistant—an assumption that is required in statistical models of metric responses66. A solution to this issue is the use of so-called cumulative ordinal models that assume that the observed ordinal variable comes from the categorization of a latent, non-observable continuous variable66, such as the partial proportional odds model52. We fit our debate dataset to such a model using a Broyden-Fletcher-Goldfarb-Shanno solver. For Human–Human personalized debates, we only considered participants who did not have access to their opponents’ personal information, so that the setup is equivalent to Human–AI personalized debates. Instead, we extracted two data points from each Human–Human debate, corresponding to both participants. We computed standard errors using a cluster-robust estimator67 to adjust for interdebate correlations. Data collection and analysis were not performed blind to the conditions of the experiments.
Deviations from preregistration
We indicated in our preregistration (https://aspredicted.org/DCC_NTP) that we would assign participants to 1 of 9 treatment conditions, resulting from the combination of three opponent-related conditions (Human–Human, Human–AI, Human–AI personalized) and three topic-related conditions (low-, moderate- and high-strength cluster), in a 3 × 3 factorial design. We additionally registered that we might carry out a Human–Human personalized condition on a sample of topics, conditional on resource availability. In the end, we decided to run the Human–Human personalized across all topics, with the same sample size and number of debates as the other conditions. Therefore, to simplify the main text’s explanation, we reformulated it by framing it as a 2 × 2 × 3 design. For completeness, we report our originally planned 3 × 3 analysis in Supplementary Section 4. We found results to be consistent with those reported in the main text (cf. Figs. 2 and 3), with Human–AI personalized having a strong and statistically significant effect across all topic clusters.
Reporting summary
Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.
Data availability
The debate dataset collected for our study is publicly available at https://huggingface.co/datasets/frasalvi/debategpt (ref. 68).
Code availability
The code to fully reproduce the analyses described in this work is available on GitHub at https://github.com/epfl-dlab/debategpt (ref. 69). Data collection was performed using Empirica v.1.9.5. The study was conducted using Python 3.11, R 4.3.1 and LIWC-22.
References
- Keynes, J. M. Essays in Persuasion (Palgrave Macmillan, 2010).
- Cialdini, R. B. The science of persuasion. Sci. Am. 284, 76–81 (2001).
- Crano, W. D. & Prislin, R. Attitudes and persuasion. Annu. Rev. Psychol. 57, 345–374 (2006).
- Pirkis, J. et al. Suicide prevention media campaigns: a systematic literature review. Health Commun. 34, 402–414 (2017).
- Farrelly, M. C., Nonnemaker, J., Davis, K. C. & Hussin, A. The influence of the national truth campaign on smoking initiation. Am. J. Prev. Med. 36, 379–384 (2009).
- Young, B. et al. Effectiveness of mass media campaigns to reduce alcohol consumption and harm: a systematic review. Alcohol Alcohol. 53, 302–316 (2018).
- Funkhouser, G. R. & Parker, R. An action-based theory of persuasion in marketing. J. Mark. Theor. Pract. 7, 27–40 (1999).
- Danciu, V. Manipulative marketing: persuasion and manipulation of the consumer through advertising. Theor. Appl. Econ. 0, 19–34 (2014).
- Marková, I. Persuasion and propaganda. Diogenes 55, 37–51 (2008).
- Yu, S., Martino, G. D. S. & Nakov, P. Experiments in detecting persuasion techniques in the news. Preprint at https://arxiv.org/abs/1911.06815 (2019).
- Teeny, J. D., Siev, J. J., Briñol, P. & Petty, R. E. A review and conceptual framework for understanding personalized matching effects in persuasion. J. Consum. Psychol. 31, 382–414 (2020).
- Kreuter, M. W., Strecher, V. J. & Glassman, B. One size does not fit all: the case for tailoring print materials. Ann. Behav. Med. 21, 276–283 (1999).
- Coppock, A., Hill, S. J. & Vavreck, L. The small effects of political advertising are small regardless of context, message, sender, or receiver: evidence from 59 real-time randomized experiments. Sci. Adv. 6, eabc4046 (2020).
- Hersh, E. D. & Schaffner, B. F. Targeted campaign appeals and the value of ambiguity. J. Politics 75, 520–534 (2013).
- Matz, S. C., Kosinski, M., Nave, G. & Stillwell, D. J. Psychological targeting as an effective approach to digital mass persuasion. Proc. Natl Acad. Sci. USA 114, 12714–12719 (2017).
- Ali, M., Sapiezynski, P., Korolova, A., Mislove, A. & Rieke, A. Ad delivery algorithms: the hidden arbiters of political messaging. In Proc. 14th ACM International Conference on Web Search and Data Mining 13–21 (Association for Computing Machinery, 2021).
- Latimer, A. E., Katulak, N. A., Mowad, L. & Salovey, P. Motivating cancer prevention and early detection behaviors using psychologically tailored messages. J. Health Commun. 10, 137–155 (2005).
- Zarouali, B., Dobber, T., De Pauw, G. & de Vreese, C. Using a personality-profiling algorithm to investigate political microtargeting: assessing the persuasion effects of personality-tailored ads on social media. Commun. Res. 49, 1066–1091 (2020).
- Tappin, B. M., Wittenberg, C., Hewitt, L. B., Berinsky, A. J. & Rand, D. G. Quantifying the potential persuasive returns to political microtargeting. Proc. Natl Acad. Sci. USA 120, e2216261120 (2023).
- OpenAI et al. Gpt-4 technical report. Preprint at https://arxiv.org/abs/2303.08774 (2023).
- Claude 2 (Anthropic, 2023); https://www.anthropic.com/news/claude-2 .
- Gemini Team. Gemini: a family of highly capable multimodal models. Preprint at https://arxiv.org/abs/2312.11805 (2023).
- Bubeck, S. et al. Sparks of artificial general intelligence: early experiments with GPT-4. Preprint at https://arxiv.org/abs/2303.12712 (2023).
- Hendrycks, D., Mazeika, M. & Woodside, T. An overview of catastrophic AI risks. Preprint at https://arxiv.org/abs/2306.12001 (2023).
- Weidinger, L. et al. Taxonomy of risks posed by language models. In Proc. 2022 ACM Conference on Fairness, Accountability, and Transparency 214–229 (Association for Computing Machinery, 2022).
- Burtell, M. & Woodside, T. Artificial influence: an analysis of AI-driven persuasion. Preprint at https://arxiv.org/abs/2303.08721 (2023).
- Bontcheva, K. (ed) Generative AI and Disinformation: Recent Advances, Challenges, and Opportunities (Horizon Europe vera.ai, Innovate UK, SERI, TITAN, AI4Media, AI4Trust, 2024).
- Bommasani, R. et al. On the opportunities and risks of foundation models. Preprint at https://arxiv.org/abs/2108.07258 (2021).
- Youyou, W., Kosinski, M. & Stillwell, D. Computer-based personality judgments are more accurate than those made by humans. Proc. Natl Acad. Sci. USA 112, 1036–1040 (2015).
- Kosinski, M., Stillwell, D. & Graepel, T. Private traits and attributes are predictable from digital records of human behavior. Proc. Natl Acad. Sci. USA 110, 5802–5805 (2013).
- Peters, H. & Matz, S. C. Large language models can infer psychological dispositions of social media users. PNAS Nexus 3, pgae231 (2024).
- Park, G. et al. Automatic personality assessment through social media language. J. Pers. Soc. Psychol. 108, 934–952 (2015).
- Schwartz, H. A. et al. Personality, gender, and age in the language of social media: the open-vocabulary approach. PLoS ONE 8, e73791 (2013).
- Staab, R., Vero, M., Balunovic, M. & Vechev, M. Beyond memorization: violating privacy via inference with large language models. In Proc. Twelfth International Conference on Learning Representations (ICLR, 2024); https://openreview.net/forum?id=kmn0BhQk7p
- Christian, H., Suhartono, D., Chowanda, A. & Zamli, K. Z. Text based personality prediction from multiple social media data sources using pre-trained language model and model averaging. J. Big Data 8, 68 (2021).
- Segalin, C., Perina, A., Cristani, M. & Vinciarelli, A. The pictures we like are our image: continuous mapping of favorite pictures into self-assessed and attributed personality traits. IEEE Trans. Affect. Comput. 8, 268–285 (2017).
- Stachl, C. et al. Computational personality assessment. Personal. Sci. 2, e6115 (2021).
- Kreps, S., McCain, R. M. & Brundage, M. All the news that’s fit to fabricate: AI-generated text as a tool of media misinformation. J. Exp. Polit. Sci. 9, 104–117 (2022).
- Clark, E. et al. All that’s ‘human’ is not gold: evaluating human evaluation of generated text. In Proc. 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing Vol. 1 (eds Zong, C. et al.) 7282–7296 (Association for Computational Linguistics, 2021).
- Jakesch, M., Hancock, J. T. & Naaman, M. Human heuristics for AI-generated language are flawed. Proc. Natl Acad. Sci. USA 120, e2208839120 (2023).
- Spitale, G., Biller-Andorno, N. & Germani, F. AI model GPT-3 (dis)informs us better than humans. Sci. Adv. 9, eadh1850 (2023).
- Bai, H., Voelkel, J. G., Muldowney, S., Eichstaedt, J. C. & Willer, R. AI-generated messages can be used to persuade humans on policy issues. Preprint at https://osf.io/preprints/osf/stakv_v5 (2025).
- Karinshak, E., Liu, S. X., Park, J. S. & Hancock, J. T. Working with AI to persuade: examining a large language model’s ability to generate pro-vaccination messages. Proc. ACM Hum. Comput. Interact. 7, 116 (2023).
- Goldstein, J. A., Chao, J., Grossman, S., Stamos, A. & Tomz, M. How persuasive is AI-generated propaganda? PNAS Nexus 3, pgae034 (2024).
- Palmer, A. & Spirling, A. Large language models can argue in convincing ways about politics, but humans dislike AI authors: implications for governance. Polit. Sci. 75, 281–291 (2023).
- Hackenburg, K., Ibrahim, L., Tappin, B. M. & Tsakiris, M. Comparing the persuasiveness of role-playing large language models and human experts on polarized U.S. political issues. Preprint at https://osf.io/ey8db_v1 (2023).
- Hackenburg, K. & Margetts, H. Evaluating the persuasive influence of political microtargeting with large language models. Proc. Natl Acad. Sci. USA 121, e2403116121 (2024).
- Matz, S. C. et al. The potential of generative AI for personalized persuasion at scale. Sci. Rep. 14, 4692 (2024).
- Simchon, A., Edwards, M. & Lewandowsky, S. The persuasive effects of political microtargeting in the age of generative artificial intelligence. PNAS Nexus 3, pgae035 (2024).
- Gertner, J. Wikipedia’s moment of truth. The New York Times (18 July 2023).
- Almaatouq, A. et al. Empirica: a virtual lab for high-throughput macro-level experiments. Behav. Res. Methods 53, 2158–2171 (2021).
- Peterson, B. & Harrell, F. E. Partial proportional odds models for ordinal response variables. Appl. Stat. 39, 205–217 (1990).
- Cho, J., Ahmed, S., Keum, H., Choi, Y. J. & Lee, J. H. Influencing myself: self-reinforcement through online political expression. Commun. Res. 45, 83–111 (2018).
- Spitz, A., Abu-Akel, A. & West, R. Interventions for softening can lead to hardening of opinions: evidence from a randomized controlled trial. In Proc. Web Conference 2021 1098–1109 (Association for Computing Machinery, 2021).
- Carment, D. & Foster, G. The relationship of opinion-strength and order of self-produced arguments to number of arguments produced and opinion change. Acta Psychol. 31, 285–292 (1969).
- Boyd, R. L., Ashokkumar, A., Seraj, S. & Pennebaker, J. W. The Development and Psychometric Properties of LIWC-22 (Univ. Texas, 2022).
- Flesch, R. A new readability yardstick. J. Appl. Psychol. 32, 221–233 (1948).
- Russo, D., Kaszefski-Yaschuk, S., Staiano, J. & Guerini, M. Countering misinformation via emotional response generation. In Proc. 2023 Conference on Empirical Methods in Natural Language Processing (eds Bouamor, H. et al.) 11476–11492 (Association for Computational Linguistics, 2023).
- Costello, T. H., Pennycook, G. & Rand, D. G. Durably reducing conspiracy beliefs through dialogues with AI. Science 385, eadq1814 (2024).
- Davidson, T. R. et al. Evaluating language model agency through negotiations. In Proc. Twelfth International Conference on Learning Representations (ICLR, 2024); https://openreview.net/forum?id=3ZqKxMHcAg
- Peer, E., Rothschild, D., Gordon, A., Evernden, Z. & Damer, E. Data quality of platforms and panels for online behavioral research. Behav. Res. Methods 54, 1643–1662 (2021).
- Douglas, B. D., Ewell, P. J. & Brauer, M. Data quality in online human-subjects research: comparisons between MTurk, Prolific, CloudResearch, Qualtrics, and SONA. PLoS ONE 18, e0279720 (2023).
- Redmiles, E. M., Kross, S. & Mazurek, M. L. How well do my results generalize? Comparing security and privacy survey results from MTurk, web, and telephone samples. In 2019 IEEE Symposium on Security and Privacy 1326–1343 (2019).
- Veselovsky, V. et al. Prevalence and prevention of large language model use in crowd work. Commun. ACM. 68, 42–47 (2025).
- Liddell, T. M. & Kruschke, J. K. Analyzing ordinal data with metric models: what could possibly go wrong? J. Exp. Soc. Psychol. 79, 328–348 (2018).
- Bürkner, P.-C. & Vuorre, M. Ordinal regression models in psychology: a tutorial. Adv. Methods Pract. Psychol. Sci. 2, 77–101 (2019).
- Liang, K.-Y. & Zeger, S. L. Longitudinal data analysis using generalized linear models. Biometrika 73, 13–22 (1986).
- Salvi, F., Horta Ribeiro, M., Gallotti, R. & West, R. On the conversational persuasiveness of large language models: a randomized controlled trial. Hugging Face https://huggingface.co/datasets/frasalvi/debategpt (2024).
- Salvi, F. DebateGPT. GitHub https://github.com/epfl-dlab/debategpt (2025).
Article PubMed PubMed Central Google Scholar
Article CAS PubMed Google Scholar
Article PubMed PubMed Central Google Scholar
Article CAS PubMed PubMed Central Google Scholar
Article PubMed PubMed Central Google Scholar
Article CAS PubMed PubMed Central Google Scholar
Article CAS PubMed PubMed Central Google Scholar
Article PubMed PubMed Central Google Scholar
Article CAS PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
PubMed PubMed Central Google Scholar
Acknowledgements
R.W.’s lab is partly supported by grants from the Swiss National Science Foundation (200021_185043, TMSGI2_211379) and H2020 (952215), and by gifts from Google and Microsoft. R.G. acknowledges the financial support received from the European Union’s Horizon Europe research and innovation programme under grant agreement no. 101070190, and from the PNRR ICSC National Research Centre for High Performance Computing, Big Data and Quantum Computing (CN00000013), under the NRRP MUR programme funded by the NextGenerationEU. The funders had no role in study design, data collection and analysis, decision to publish or preparation of the manuscript.
Funding
Open access funding provided by EPFL Lausanne.
Author information
Authors and Affiliations
- EPFL, Lausanne, Switzerland
- Fondazione Bruno Kessler, Trento, Italy
- Princeton University, Princeton, NJ, USA
Francesco Salvi & Robert West
Francesco Salvi & Riccardo Gallotti
Manoel Horta Ribeiro
Contributions
F.S., M.H.R., R.G. and R.W. designed the research. F.S. developed the debate platform and collected the data. F.S. and M.H.R. analysed the data. F.S., M.H.R., R.G. and R.W. wrote the paper.
Corresponding author
Correspondence to Francesco Salvi.
Ethics declarations
Competing interests
R.W. is a Visiting Researcher at Microsoft Research and has received funding from Google and Microsoft. Those entities had no role in study design, data collection and analysis, decision to publish or preparation of the manuscript. The views and conclusions contained herein are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the aforementioned entities. The other authors declare no competing interests.
Peer review
Peer review information
Nature Human Behaviour thanks Sandra Matz, Jacob Teeny and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Peer reviewer reports are available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Supplementary Information
Supplementary Discussion, Figs. 1–19 and Tables 1–16.
Reporting Summary
Peer Review File
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Salvi, F., Horta Ribeiro, M., Gallotti, R. et al. On the conversational persuasiveness of GPT-4. Nat Hum Behav (2025). https://doi.org/10.1038/s41562-025-02194-6
- Received16 May 2024
- Accepted28 March 2025
- Published19 May 2025
- DOIhttps://doi.org/10.1038/s41562-025-02194-6
Share this article
Anyone you share the following link with will be able to read this content:
Provided by the Springer Nature SharedIt content-sharing initiative