Could AI one day convince world leaders to make dangerous decisions? OpenAI, the creator of ChatGPT, is keeping tabs on this possibility. While AI models are getting better at crafting persuasive arguments, they’re not quite "superhuman" yet. But the company isn’t taking any chances.
OpenAI has been testing its models against Reddit’s r/ChangeMyView, a forum where users debate and change opinions. They compare AI-generated responses to human replies, using a five-point scale to measure persuasiveness. The latest AI models are now more persuasive than humans in about 82% of cases. While that’s impressive, it’s still far from the 95th percentile OpenAI considers “superhuman.”
The risks of AI persuasion are real. Even human-level persuasive writing could fuel biased journalism, sway elections, or bolster scams. OpenAI has labeled this a “Medium” risk in its Preparedness Framework. They’re already monitoring for AI-powered influence operations and have set rules to block AI from handling political persuasion tasks.
What’s the big deal? Well, crafting a strong argument takes time and effort for humans. For AI, it’s nearly instant and cost-free. This could lead to a flood of AI-generated content, making it easier to manipulate public opinion on a large scale.
For now, OpenAI is staying vigilant. While we’re not yet at the point where AI can hypnotize world leaders into bad decisions, the company is making sure it’s ready if that day ever comes.