ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2407.03234
  4. Cited By
Self-Evaluation as a Defense Against Adversarial Attacks on LLMs

Self-Evaluation as a Defense Against Adversarial Attacks on LLMs

3 July 2024
Hannah Brown
Leon Lin
Kenji Kawaguchi
Michael Shieh
    AAML
ArXivPDFHTML

Papers citing "Self-Evaluation as a Defense Against Adversarial Attacks on LLMs"

2 / 2 papers shown
Title
ETA: Evaluating Then Aligning Safety of Vision Language Models at Inference Time
ETA: Evaluating Then Aligning Safety of Vision Language Models at Inference Time
Yi Ding
Bolian Li
Ruqi Zhang
MLLM
67
5
0
09 Oct 2024
Self-Evaluation Guided Beam Search for Reasoning
Self-Evaluation Guided Beam Search for Reasoning
Yuxi Xie
Kenji Kawaguchi
Yiran Zhao
Xu Zhao
MingSung Kan
Junxian He
Qizhe Xie
LRM
164
128
0
01 May 2023
1