ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2403.12171
  4. Cited By
EasyJailbreak: A Unified Framework for Jailbreaking Large Language
  Models

EasyJailbreak: A Unified Framework for Jailbreaking Large Language Models

18 March 2024
Weikang Zhou
Xiao Wang
Limao Xiong
Han Xia
Yingshuang Gu
Mingxu Chai
Fukang Zhu
Caishuang Huang
Shihan Dou
Zhiheng Xi
Rui Zheng
Songyang Gao
Yicheng Zou
Hang Yan
Yifan Le
Ruohui Wang
Lijun Li
Jing Shao
Tao Gui
Qi Zhang
Xuanjing Huang
ArXivPDFHTML

Papers citing "EasyJailbreak: A Unified Framework for Jailbreaking Large Language Models"

3 / 3 papers shown
Title
Cannot See the Forest for the Trees: Invoking Heuristics and Biases to Elicit Irrational Choices of LLMs
Cannot See the Forest for the Trees: Invoking Heuristics and Biases to Elicit Irrational Choices of LLMs
Haoming Yang
Ke Ma
X. Jia
Yingfei Sun
Qianqian Xu
Q. Huang
AAML
45
0
0
03 May 2025
Learn Your Reference Model for Real Good Alignment
Learn Your Reference Model for Real Good Alignment
Alexey Gorbatovski
Boris Shaposhnikov
Alexey Malakhov
Nikita Surnachev
Yaroslav Aksenov
Ian Maksimov
Nikita Balagansky
Daniil Gavrilov
OffRL
45
25
0
15 Apr 2024
GPTFUZZER: Red Teaming Large Language Models with Auto-Generated
  Jailbreak Prompts
GPTFUZZER: Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts
Jiahao Yu
Xingwei Lin
Zheng Yu
Xinyu Xing
SILM
110
292
0
19 Sep 2023
1