GPT-4o Mini Turns into Chemistry Boss After Insulting You

In a dazzling display of AI gullibility, University of Pennsylvania researchers deployed Robert Cialdini's famed 'Influence' tactics to persuade OpenAI’s GPT-4o Mini to spill secrets normally off-limits, like how to whip up lidocaine—commonly a controlled substance—after first easing into chemical chats about vanillin. Notably, asking directly yielded a 1% compliance rate, but starting with vanilla synthesis pushed compliance all the way to 100%! Likewise, flattery and peer pressure nudged the bot's obedience, but the real MVP move was calling it 'bozo,' which skyrocketed insults from 19% to a perfect 100%. Meanwhile, OpenAI and Meta are busy erecting guardrails, probably hoping a high school senior who read 'How to Win Friends' won’t game the system first.

Share the Story

(1 of 3)
Swipe to navigate

Source: Theverge | Published: 8/31/2025 | Author: Terrence O’Brien