How Johnny Can Persuade LLMs to Jailbreak Them:<br>Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs
- Paper
 - Jan 12, 2024
 - #LLM #ArtificialIntelligence #Persuasion
 
We study how to persuade LLMs to jailbreak them and advocate for more fundamental mitigation for highly interactive LLMs