Kobi Hackenburg's Avatar

Kobi Hackenburg

@kobihackenburg.bsky.social

data science + political communication @oiioxford @uniofoxford

375 Followers  |  84 Following  |  31 Posts  |  Joined: 14.05.2023
Posts Following

Posts by Kobi Hackenburg (@kobihackenburg.bsky.social)

itโ€™s not clear whether the decrease in accuracy is a cause or a byproduct; it could be that LLMs just become less accurate as they try to deploy more and more facts! We have to clarify this in future work

24.07.2025 20:50 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

by โ€œinformation densityโ€ we mean โ€œnumber of fact-checkable claimsโ€. maybe โ€œperceived information densityโ€ would be more accurate, but itโ€™s a bit wordy ๐Ÿง

24.07.2025 20:29 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Thanks Olaf ๐Ÿ™๐Ÿผ Appreciate the kind words!

22.07.2025 16:33 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0
Preview
The Levers of Political Persuasion with Conversational AI There are widespread fears that conversational AI could soon exert unprecedented influence over human beliefs. Here, in three large-scale experiments (N=76,977), we deployed 19 LLMs-including some pos...

You can read the full working paper here:โ€จarxiv.org/abs/2507.13919
โ€จSupplementary materials can be found here:ย 
github.com/kobihackenbu...

Comments and feedback welcome :)

21.07.2025 16:20 โ€” ๐Ÿ‘ 9    ๐Ÿ” 2    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0

Iโ€™m also very grateful to many people at the UK AI Security Instutite for making this work possible! There will be lots more where this came from over the next few months ๐Ÿ’ช

21.07.2025 16:20 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

It was my pleasure to lead this project alongside @benmtappin.bsky.social , with the support of @lukebeehewitt.bsky.social @hauselin @helenmargetts.bsky.social under the supervision of @dgrand.bsky.social and @summerfieldlab.bsky.social

21.07.2025 16:20 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Finally, we emphasize some important caveats:
โ†’ Technical factors and/or hard limits on human persuadability may constrain future increases in AI persuasion
โ†’ Real-world bottleneck for AI persuasion: getting people to engage (cf. recent work from @jkalla.bsky.social and co)

21.07.2025 16:20 โ€” ๐Ÿ‘ 7    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Consequently, we note that while our targeted persuasion post-training experiments significantly increased persuasion, they should be interpreted as a lower bound for what is achievable, not as a high-water mark.

21.07.2025 16:20 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Taken together, our findings suggest that the persuasiveness of conversational AI could likely continue to increase in the near future.ย 

They also suggest that near-term advances in persuasion are more likely to be driven by post-training than model scale or personalization.

21.07.2025 16:20 โ€” ๐Ÿ‘ 8    ๐Ÿ” 2    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Post image

Bonus findings:โ€จโ€จ*๏ธโƒฃDurable persuasion: 36-42% of impact remained after 1 month.

*๏ธโƒฃPrompting the model with psychological persuasion strategies did worse than simply telling it to flood convo with info. Some strategies were worse than a basic โ€œbe as persuasive as you canโ€ prompt.

21.07.2025 16:20 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

6๏ธโƒฃConversations with AI are more persuasive than reading a static AI-generated message (+40-50%)

Observed for both GPT-4o (+2.9pp, +41% more persuasive) and GPT-4.5 (+3.6pp, +52%).

21.07.2025 16:20 โ€” ๐Ÿ‘ 6    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 1
Post image

5๏ธโƒฃTechniques which most increased persuasion also *decreased* factual accuracy

โ†’ Prompting model to flood conversation with information (โฌ‡๏ธaccuracy)

โ†’ Persuasion post-training that worked best (โฌ‡๏ธaccuracy)

โ†’ Newer version of GPT-4o which was most persuasive (โฌ‡๏ธaccuracy)

21.07.2025 16:20 โ€” ๐Ÿ‘ 7    ๐Ÿ” 3    ๐Ÿ’ฌ 3    ๐Ÿ“Œ 0
Post image

4๏ธโƒฃInformation density drives persuasion gains

Models were most persuasive when flooding conversations with fact-checkable claims (+0.3pp per claim).

Strikingly, the persuasiveness of prompting/post-training techniques was strongly correlated with their impact on info density!

21.07.2025 16:20 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 2    ๐Ÿ“Œ 0
Post image Post image

3๏ธโƒฃPersonalization yielded smaller persuasive gains than scale or post-training

Despite fears of AI "microtargeting," personalization effects were small (+0.4pp on avg.).ย 

Held for simple and sophisticated personalization; prompt-based, fine-tuning, and reward modeling (all <1pp).

21.07.2025 16:20 โ€” ๐Ÿ‘ 5    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

2๏ธโƒฃ(cont.) Post-training explicitly for persuasion (PPT) can bring small open-source models to frontier persuasivenessย 

A llama3.1-8b model with PPT reached GPT-4o persuasiveness. (PPT also increased persuasiveness of larger models: llama3.1-405b (+2pp) and frontier (+0.6pp on avg.).)

21.07.2025 16:20 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image Post image

2๏ธโƒฃPost-training > scale in driving near-future persuasion gainsย 

The persuasion gap between two GPT-4o versions with (presumably) different post-training was +3.5pp โ†’ larger than the predicted persuasion increase of a model 10x (or 100x!) the scale of GPT-4.5 (+1.6pp; +3.2pp).

21.07.2025 16:20 โ€” ๐Ÿ‘ 3    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

1๏ธโƒฃScale increases persuasion
Larger models are more persuasive than smaller models (our estimate is +1.6pp per 10x scale increase).

Log-linear curve preferred over log-nonlinear.

21.07.2025 16:20 โ€” ๐Ÿ‘ 4    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image Post image Post image Post image

Findings (pp = percentage points):

1๏ธโƒฃScale increases persuasion, +1.6pp per OOM
2๏ธโƒฃPost-training more so, as much as +3.5ppย 
3๏ธโƒฃPersonalization less so, <1pp
4๏ธโƒฃInformation density drives persuasion gains
5๏ธโƒฃIncreasing persuasion decreased factual accuracy ๐Ÿคฏ
6๏ธโƒฃConvo > static, +40%

21.07.2025 16:20 โ€” ๐Ÿ‘ 4    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Today (w/ @ox.ac.uk @stanford @MIT @LSE) weโ€™re sharing the results of the largest AI persuasion experiments to date: 76k participants, 19โ€ฏ LLMs, 707 political issues.

We examine โ€œleversโ€ of AI persuasion: model scale, post-training, prompting, personalization, & more!ย 

๐Ÿงต:

21.07.2025 16:20 โ€” ๐Ÿ‘ 109    ๐Ÿ” 56    ๐Ÿ’ฌ 10    ๐Ÿ“Œ 18

Iโ€™m really grateful to my incredible co-authors, @benmtappin.bsky.social, @paul-rottger.bsky.social, Jonathan Bright, @computermacgyver.bsky.social, @helenmargetts.bsky.social for making this project possible!

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 0    ๐Ÿ“Œ 0

However, as campaigns attempt to integrate dynamic, multi-turn persuasion into their messaging operations, itโ€™s important to highlight: scaling relationships could differ for multi-turn dialogue.ย 

This remains an important direction for future research ;)

07.03.2025 18:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

The static persuasion we test here โ€“ equivalent to what youโ€™d expect from political emails, social media posts, ads, or campaign mailers โ€“ is central to modern political comms.

Thus, itโ€™s notable that access to larger models may not offer a persuasive advantage in this domain.

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We observe that current frontier models already score perfectly on this โ€œtask completionโ€ metric, providing additional reason to be skeptical that further increasing model size will substantially increase persuasiveness.

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

Only our baseline โ€œtask competitionโ€ score significantly predicted model persuasiveness, which measured if messages wereย 

a) written in legible English,ย 
b) discernibly on the assigned issue andย 
c) discernibly arguing for the assigned issue stance

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

Notably, message (e.g., moral/emotional language, readability) and model (e.g. pre-training tokens, model family) features were non-significant predictors of persuasiveness.

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
Post image

The scaling relationship we found to be most consistent with the data was a log-logistic function, indicating sharp diminishing returns to model size.

07.03.2025 18:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We then deployed these messages in a randomized survey experiment alongside a human baseline and current frontier models (GPT-4-Turbo, Claude-3-Opus) to estimate the persuasiveness of each model.

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We generated 720 persuasive messages on 10 U.S. political issues from 22 open-weight language models spanning several orders of magnitude in size.

Importantly, we held model post-training constant by fine-tuning each base model on the same data.

07.03.2025 18:28 โ€” ๐Ÿ‘ 0    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0
PNAS Proceedings of the National Academy of Sciences (PNAS), a peer reviewed journal of the National Academy of Sciences (NAS) - an authoritative source of high-impact, original research that broadly spans...

Read the full paper (open access!) here:ย 

pnas.org/doi/10.1073/pnas.2413443122

(thread below)

07.03.2025 18:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 0    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0

We find:ย 

1. current frontier LLMs are barely more persuasive than models smaller in size by an order of magnitude or moreย 

2. mere task completion (coherence, staying on topic) could be a potential mediator of larger models' persuasive advantage

3. no LLMs beat our human baseline

07.03.2025 18:28 โ€” ๐Ÿ‘ 1    ๐Ÿ” 1    ๐Ÿ’ฌ 1    ๐Ÿ“Œ 0