ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2406.13662
  4. Cited By
Jailbreaking Large Language Models Through Alignment Vulnerabilities in Out-of-Distribution Settings

Jailbreaking Large Language Models Through Alignment Vulnerabilities in Out-of-Distribution Settings

19 June 2024
Yue Huang
Jingyu Tang
Dongping Chen
Bingda Tang
Yao Wan
Lichao Sun
Philip S. Yu
Xiangliang Zhang
    AAML
ArXivPDFHTML

Papers citing "Jailbreaking Large Language Models Through Alignment Vulnerabilities in Out-of-Distribution Settings"

1 / 1 papers shown
Title
Poisoning Language Models During Instruction Tuning
Poisoning Language Models During Instruction Tuning
Alexander Wan
Eric Wallace
Sheng Shen
Dan Klein
SILM
90
185
0
01 May 2023
1