Ali Dehghantanha

h-index: 2 6 citations 6 papers (total)

Papers in Database (1)

attack CIKM Oct 24, 2025 · Oct 2025

Uncovering the Persuasive Fingerprint of LLMs in Jailbreaking Attacks

Havva Alizadeh Noughabi, Julien Serbanescu, Fattane Zarrinkalam et al. · University of Guelph

Exploits social-science persuasion theories to craft natural-language jailbreak prompts that bypass LLM alignment safeguards

Prompt Injection nlp
PDF Code