In a study, "On the Conversational Persuasiveness of Large Language Models: A Randomized Controlled Trial", researchers Francesco Salvi, Manoel Horta Ribeiro, Riccardo Gallotti, and Robert West shed light on the persuasive capabilities of large language models (LLMs) in direct conversations with human counterparts. The results, which were pre-registered, provide compelling evidence that AI-driven persuasion, particularly when enhanced by personalization, can significantly outperform human persuasion.

On the Conversational Persuasiveness of Large Language Models: A Randomized Controlled Trial
The development and popularization of large language models (LLMs) have raised concerns that they will be used to create tailor-made, convincing arguments to push false or misleading narratives online. Early work has found that language models can generate content perceived as at least on par and often more persuasive than human-written messages. However, there is still limited knowledge about LLMs’ persuasive capabilities in direct conversations with human counterparts and how personalization can improve their performance. In this pre-registered study, we analyze the effect of AI-driven persuasion in a controlled, harmless setting. We create a web-based platform where participants engage in short, multiple-round debates with a live opponent. Each participant is randomly assigned to one of four treatment conditions, corresponding to a two-by-two factorial design: (1) Games are either played between two humans or between a human and an LLM; (2) Personalization might or might not be enabled, granting one of the two players access to basic sociodemographic information about their opponent. We found that participants who debated GPT-4 with access to their personal information had 81.7% (p < 0.01; N=820 unique participants) higher odds of increased agreement with their opponents compared to participants who debated humans. Without personalization, GPT-4 still outperforms humans, but the effect is lower and statistically non-significant (p=0.31). Overall, our results suggest that concerns around personalization are meaningful and have important implications for the governance of social media and the design of new online environments.

Table of Contents

  1. Discussion & Insights
  2. Key Takeaways
  3. The Experimental Setup: A Web-Based Debating Platform
  4. Measuring Persuasion: Agreement Shifts and Opinion Fluidity
  5. The Power of Personalization: AI's Edge Over Humans
  6. Textual Analysis: Unraveling the Language of Persuasion
  7. Implications and Future Directions
  8. Mastering Personalization Techniques in Prompt Engineering
  9. A Prompt Engineering Framework for Persuasive AI Communication

1. Discussion & Insights

History Lesson: The Power of Persuasion

As humans, we like to think we're the masters of our own minds. But history proves otherwise. We're social creatures, easily swayed by those who can change our minds on a large scale. That's the whole basis of our society: money, rules, entire nations – they're all just ideas. Powerful ideas, sure, but still just ideas. And guess what? AI is getting terrifyingly good at manipulating those ideas.

GPT-4: The Better Debater

Studies already show that large language models like GPT-4 are scarily persuasive. In debates on hot-button issues, GPT-4 consistently changes more minds than the average human. But here's the kicker: give it a little demographic info about you, and the AI can tailor its arguments even more effectively. Basically, it wins at debate because it knows how to push your buttons.

Your Face Gives You Away

Want to make things worse? AIs are now learning to read our emotions. Companies like Hume AI are developing tech that analyzes your voice and facial expressions to figure out how you're really feeling. And when you combine that with GPT-4's argumentation skills... well, let's just say your poker face just got a lot less useful.

The Scary Possibilities

Imagine hyper-targeted ads that know exactly how to make you click 'buy'. Phone conversations with customer service bots designed to make you give up on those refunds. Political parties, law enforcement, social movements – heck, even your own HR department – using this tech to manipulate you in ways we can barely dream of. It's persuasive power on a scale that's never existed before.

2. Two Key Takeaways

1. Superior Persuasiveness of LLMs

The research highlights the inherent persuasiveness of LLMs in comparison to humans in debate settings. This is a significant insight as it showcases the advanced capabilities of models like GPT-4 in crafting arguments that resonate with audiences.

Paper: "Participants who debated GPT-4 with access to their personal information had 81.7% higher odds of increased agreement with their opponents compared to participants who debated humans."

2. Impact of Personalization

Personalization appears to be a critical factor enhancing the persuasiveness of LLMs. By tailoring arguments using participants' sociodemographic information, LLMs can significantly increase their persuasive impact.

This should not be a surprise to Prompt Engineers and students of the Prompt Engineering Institute. We've discussed this many times on the site and in our Prompt Engineering Masterclass.

Paper: "Overall, our results suggest that concerns around personalization are meaningful, showcasing how language models can out-persuade humans in online conversations through microtargeting."


3. The Experimental Setup: A Web-Based Debating Platform

The researchers developed an innovative web-based platform where participants engaged in short, multiple-round debates with a live opponent. Each participant was randomly assigned to one of four treatment conditions:

  • Human-Human: Both sides played by humans
  • Human-AI: Participants paired with an LLM (GPT-4)
  • Human-Human, personalized: Both sides human, one player has access to opponent's personal info
  • Human-AI, personalized: Participants paired with GPT-4 that has access to their personal info

This two-by-two factorial design allowed for a direct comparison of the persuasive capabilities of humans and LLMs, with and without personalization.

4. Measuring Persuasion: Agreement Shifts and Opinion Fluidity

To measure persuasion, the researchers compared participants' agreement with the debate proposition before and after the debates. Key findings include:

  • On average, LLMs significantly outperformed human participants across all topics and demographics.
  • Debating with GPT-4 with personalization resulted in an 81.7% increase in the odds of higher agreement with opponents compared to debating with humans.
  • Without personalization, GPT-4 still outperformed humans, but to a lesser extent.

The study also examined opinion fluidity, or the propensity of participants to change their minds. Factors such as topic knowledge and prior thought reduced fluidity, while topic debatableness increased it.

5. The Power of Personalization: AI's Edge Over Humans

The results highlight the remarkable ability of LLMs to exploit personal information to tailor arguments effectively. GPT-4 with personalization demonstrated far superior persuasive power compared to both non-personalized LLMs and humans with access to personal information.

This finding underscores the potential risks associated with AI-driven personalized persuasion, as malicious actors could leverage fine-grained digital traces and behavioral data to engineer highly persuasive chatbots for disinformation campaigns.

6. Textual Analysis: Unraveling the Language of Persuasion

Through a detailed textual analysis using LIWC and social dimensions, the researchers identified distinctive patterns in the language used by AI and human debaters:

  • AI players implemented logical and analytical thinking significantly more than humans.
  • Humans used more first-person singular and second-person pronouns and produced longer but easier-to-read texts.
  • AI arguments featured more factual knowledge, while human arguments displayed more appeals to similarity, expressions of support and trust, and elements of fun.

Participants correctly identified AI opponents in about 75% of cases, indicating that the writing style of LLMs has distinctive features that are relatively easy to spot.

7. Implications and Future Directions

The study's findings suggest that concerns around AI persuasion and personalization are well-founded. Online platforms and social media should seriously consider the threat of LLM-driven persuasion and implement measures to counter its spread.

Future research could focus on continuously benchmarking LLMs' persuasive capabilities, exploring different models, prompts, and settings. Additionally, the study's methodology could be extended to other contexts, such as negotiation games and open-ended conflict resolution.

While the study has some limitations, such as the predetermined structure of debates and time constraints, it provides a valuable framework for understanding the persuasive power of AI and the impact of personalization in online conversations.

8. Mastering Personalization Techniques in Prompt Engineering

As prompt engineering continues to evolve and shape the way we interact with large language models (LLMs), it's crucial for practitioners and students to understand the power of personalization. By crafting prompts that incorporate specific roles, personas, and target audiences, we can unlock the true potential of AI-driven communication.

Key Strategies for Effective Personalization

Defining a Detailed Role and Persona

    • Specify the AI's identity, background, and area of expertise
    • Establish a clear communication style and tone
    • Incorporate relevant traits and characteristics

When prompting an LLM, providing a well-defined role and persona is essential for guiding the AI's responses. By specifying details such as the AI's identity, background, expertise, and communication style, you can ensure that the generated content aligns with your desired persona.

For example, you might prompt the AI to act as a "wise, compassionate elderly grandmother named Ethel who grew up during the Great Depression." This level of specificity helps the AI generate responses that are consistent with Ethel's character, drawing upon her life experiences and warm, nurturing communication style.

Tailoring Content to a Target Audience

    • Identify the user's demographics, role, and personal details
    • Customize responses based on the user's needs and goals
    • Leverage user-specific information to create resonant content

In addition to defining a persona for the AI, it's equally important to consider the target audience. By incorporating information about the user's demographics, role, and personal details, you can prompt the LLM to generate content that is tailored to the individual.

Consider a prompt that includes details such as "The user you are conversing with is a 32-year-old male software engineer named John who lives in Seattle." By providing this context, you enable the AI to craft responses that are relevant and valuable to John's specific situation, addressing his needs and goals in a personalized manner.

The Power of Personalized AI Interactions

By combining these two personalization techniques - defining a detailed role and persona for the AI and tailoring content to a specific target audience - prompt engineers can create highly engaging and persuasive AI-driven communication.

As research continues to demonstrate the impact of personalization on the effectiveness of AI-generated content, it's clear that mastering these techniques will be essential for anyone working with LLMs. By leveraging the power of personalization, we can unlock the true potential of AI and create more meaningful, resonant interactions.

9. A Prompt Engineering Framework for Persuasive AI Communication

Based on the insights from the study "On the Conversational Persuasiveness of Large Language Models," we can develop a prompt engineering framework that leverages the power of personalization to create more persuasive and engaging AI-driven communication.

The Personalized Persuasion Prompt (PPP) Framework

This post is for paying subscribers only

Sign up now and upgrade your account to read the post and get access to the full library of posts for paying subscribers only.

Sign up now Already have an account? Sign in