Kobi Hackenburg Profile picture
Jul 21 16 tweets 6 min read Read on X
Today (w/ @UniofOxford @Stanford @MIT @LSEnews) we’re sharing the results of the largest AI persuasion experiments to date: 76k participants, 19  LLMs, 707 political issues.

We examine “levers” of AI persuasion: model scale, post-training, prompting, personalization, & more
🧵 Image
RESULTS (pp = percentage points):

1️⃣Scale increases persuasion, +1.6pp per OOM

2️⃣Post-training more so, +3.5pp 

3️⃣Personalization less so, <1pp

4️⃣Information density drives persuasion gains

5️⃣Increasing persuasion decreased factual accuracy 🤯

6️⃣Convo > static, +40% Image
Image
Image
Image
1️⃣Scale increases persuasion

Larger models are more persuasive than smaller models (our estimate is +1.6pp per 10x scale increase).

Log-linear curve preferred over log-nonlinear. Image
Image
2️⃣Post-training > scale in driving near-future persuasion gains 

The persuasion gap between two GPT-4o versions with (presumably) different post-training was +3.5pp → larger than the predicted persuasion increase of a model 10x (or 100x!) the scale of GPT-4.5 (+1.6pp; +3.2pp). Image
Image
2️⃣(cont.) Post-training explicitly for persuasion (PPT) can bring small open-source models to frontier persuasiveness 

A llama3.1-8b model with PPT reached GPT-4o persuasiveness. PPT also increased persuasiveness of larger models: llama3.1-405b (+2pp) and frontier (avg. +0.6pp) Image
3️⃣Personalization yielded smaller persuasive gains than scale or post-training

Despite fears of AI "microtargeting," personalization effects were small (+0.4pp on avg.). Held for simple and sophisticated personalization: prompting, fine-tuning, and reward modeling (all <1pp) Image
Image
4️⃣Information density drives persuasion gains

Models were most persuasive when flooding conversations with fact-checkable claims (+0.3pp per claim).

Strikingly, the persuasiveness of prompting/post-training techniques was strongly correlated with their impact on info density! Image
5️⃣Techniques which most increased persuasion also *decreased* factual accuracy

→ Prompting model to flood conversation with information (⬇️accuracy)

→ Persuasion post-training that worked best (⬇️accuracy)

→ Newer version of GPT-4o which was most persuasive (⬇️accuracy) Image
6️⃣Conversations with AI are more persuasive than reading a static AI-generated message (+40-50%)

Observed for both GPT-4o (+2.9pp, +41% more persuasive) and GPT-4.5 (+3.6pp, +52%).
Bonus stats:

*️⃣Durable persuasion: 36-42% of impact remained after 1 month.

*️⃣Prompting the model with psychological persuasion strategies did worse than simply telling it to flood convo with info. Some strategies were worse than a basic “be as persuasive as you can” prompt Image
Taken together, our findings suggest that the persuasiveness of conversational AI could likely continue to increase in the near future.

They also suggest that near-term advances in persuasion are more likely to be driven by post-training than model scale or personalization.
Consequently, we note that while our targeted persuasion post-training experiments significantly increased persuasion, they should be interpreted as a lower bound for what is achievable, not as a high-water mark.
Finally, we emphasize some important caveats:

→ Technical factors and/or hard limits on human persuadability may constrain future increases in AI persuasion

→ Real-world bottleneck for AI persuasion: getting people to engage (cf. recent work from @j_kalla and co) Image
It was my pleasure to lead this project alongside @Ben_Tappin, with the support of @lukebeehewitt @hauselin @realmeatyhuman Ed Saunders @CatherineFist @HelenMargetts under the supervision of @DG_Rand and @summerfieldlab
I’m also very grateful to many people @AISecurityInst —especially my team—for making this work possible!

There will be lots more where this came from over the next few months 👀
You can read the full working paper here:
arxiv.org/abs/2507.13919

Supplementary materials can be found here: 
github.com/kobihackenburg…

Comments and feedback welcome :)

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Kobi Hackenburg

Kobi Hackenburg Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @KobiHackenburg

Jun 21, 2024
‼️New preprint: Scaling laws for political persuasion with LLMs‼️

In a large pre-registered experiment (n=25,982), we find evidence that scaling the size of language models yields sharply diminishing persuasive returns:



1/n arxiv.org/abs/2406.14508
Image
We find that: 

➡️ current frontier models (GPT-4, Claude-3) are barely more persuasive than models smaller in size by an order of magnitude or more, and  

➡️ mere task completion (coherence, staying on topic) appears to account for larger models' persuasive advantage.
Together, these findings suggest that further scaling existing model size may not much increase the persuasiveness of static LLM-generated (political) messages! 

How we tested this:
Read 12 tweets
Jun 7, 2024
🚨New today in @PNASNews ‼️w/ @helenmargetts:

In a pre-registered experiment (n=8,587), we find little evidence that 1:1 personalization — aka microtargeting — enhances the persuasive influence of political messages generated by GPT-4. 



1/7 👇🏼 pnas.org/doi/10.1073/pn…
Image
Our findings suggest:

1️⃣Personalizing static political messages with current frontier LLMs may not offer the persuasive advantage that has been widely speculated…

2️⃣but both targeted and non-targeted messages are broadly persuasive.

How we tested this: Image
We built a custom web app to inject self-reported demographic & political data into GPT-4 prompts in real-time.

This enabled the live creation of thousands of personalized messages, tailored to persuade individual participants on an array of political issues. Image
Read 7 tweets
Jun 15, 2023
🚨 New paper! 🚨 w/ @william__brady & @manos_tsakiris:

Today in @PNASNexus we map the moral language of 39 U.S. presidential candidates to show how they're connected / differentiated by their use of moral rhetoric.
ow.ly/8Zwu50OO5QA

A [visual] thread on findings: Image
Main takeaways:

1) @TheDemocrats & @GOP
candidates use sharply divergent moral vocabularies

2) Candidates can separate themselves from the rhetorical norms of their party by using unique moral language, but rarely do
Taken together with prior research (@RobbWiller), this work empirically shows that the moral rhetoric employed by recent candidates is probably un-persuasive across party lines & likely exacerbates polarization, particularly in online social networks.
Read 13 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(