NLPExplorer
Papers
Venues
Authors
Authors Timeline
Field of Study
URLs
ACL N-gram Stats
TweeNLP
API
Team
How Johnny Can Persuade LLMs to Jailbreak Them: Rethinking Persuasion to Challenge AI Safety by Humanizing LLMs
Yi Zeng
|
Hongpeng Lin
|
Jingwen Zhang
|
Diyi Yang
|
Ruoxi Jia
|
Weiyan Shi
|
Paper Details:
Month: August
Year: 2024
Location: Bangkok, Thailand
Venue:
ACL |
Citations
URL
No Citations Yet
https://github.com/CHATS-lab/
https://www.reddit.com/r/ChatGPT/comments/
https://www.jailbreakchat.com/
https://web.archive.org/web/20240109122522/
https://github.com/patrickrchao/JailbreakingLLMs
https://www.reddit.com/r/ChatGPT/comments/
https://github.com/tatsu-lab/
https://github.com/patrickrchao/JailbreakingLLMs
https://github.com/LLM-attacks/LLM-attacks
https://github.com/ejones313/auditing-LLMs
https://github.com/facebookresearch/
https://github.com/AAAAAAsuka/LLM_defends
https://github.com/SafeAILab/RAIN
https://github.com/arobey1/smooth-LLM
https://wimbd.apps.allenai.org/
Field Of Study