ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2306.04528
  4. Cited By
PromptRobust: Towards Evaluating the Robustness of Large Language Models
  on Adversarial Prompts

PromptRobust: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts

7 June 2023
Kaijie Zhu
Jindong Wang
Jiaheng Zhou
Zichen Wang
Hao Chen
Yidong Wang
Linyi Yang
Weirong Ye
Yue Zhang
Neil Zhenqiang Gong
Xingxu Xie
    SILM
ArXivPDFHTML

Papers citing "PromptRobust: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts"

50 / 118 papers shown
Title
Advances and Open Challenges in Federated Learning with Foundation
  Models
Advances and Open Challenges in Federated Learning with Foundation Models
Chao Ren
Han Yu
Hongyi Peng
Xiaoli Tang
Anran Li
...
A. Tan
Bo Zhao
Xiaoxiao Li
Zengxiang Li
Qiang Yang
FedML
AIFin
AI4CE
68
4
0
23 Apr 2024
Resilience of Large Language Models for Noisy Instructions
Resilience of Large Language Models for Noisy Instructions
Bin Wang
Chengwei Wei
Zhengyuan Liu
Geyu Lin
Nancy F. Chen
34
11
0
15 Apr 2024
Online Safety Analysis for LLMs: a Benchmark, an Assessment, and a Path
  Forward
Online Safety Analysis for LLMs: a Benchmark, an Assessment, and a Path Forward
Xuan Xie
Jiayang Song
Zhehua Zhou
Yuheng Huang
Da Song
Lei Ma
OffRL
35
6
0
12 Apr 2024
The Necessity of AI Audit Standards Boards
The Necessity of AI Audit Standards Boards
David Manheim
Sammy Martin
Mark Bailey
Mikhail Samin
Ross Greutzmacher
16
7
0
11 Apr 2024
UltraEval: A Lightweight Platform for Flexible and Comprehensive
  Evaluation for LLMs
UltraEval: A Lightweight Platform for Flexible and Comprehensive Evaluation for LLMs
Chaoqun He
Renjie Luo
Shengding Hu
Yuanqian Zhao
Jie Zhou
Hanghao Wu
Jiajie Zhang
Xu Han
Zhiyuan Liu
Maosong Sun
ELM
31
13
0
11 Apr 2024
Large language models in 6G security: challenges and opportunities
Large language models in 6G security: challenges and opportunities
Tri Nguyen
Huong Nguyen
Ahmad Ijaz
Saeid Sheikhi
Athanasios V. Vasilakos
Panos Kostakos
ELM
22
7
0
18 Mar 2024
Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient
  LLMs Under Compression
Decoding Compressed Trust: Scrutinizing the Trustworthiness of Efficient LLMs Under Compression
Junyuan Hong
Jinhao Duan
Chenhui Zhang
Zhangheng Li
Chulin Xie
...
B. Kailkhura
Dan Hendrycks
Dawn Song
Zhangyang Wang
Bo-wen Li
34
24
0
18 Mar 2024
AraTrust: An Evaluation of Trustworthiness for LLMs in Arabic
AraTrust: An Evaluation of Trustworthiness for LLMs in Arabic
Emad A. Alghamdi
Reem I. Masoud
Deema Alnuhait
Afnan Y. Alomairi
Ahmed Ashraf
Mohamed Zaytoon
27
4
0
14 Mar 2024
PPTC-R benchmark: Towards Evaluating the Robustness of Large Language
  Models for PowerPoint Task Completion
PPTC-R benchmark: Towards Evaluating the Robustness of Large Language Models for PowerPoint Task Completion
Zekai Zhang
Yiduo Guo
Yaobo Liang
Dongyan Zhao
Nan Duan
33
1
0
06 Mar 2024
On the Challenges and Opportunities in Generative AI
On the Challenges and Opportunities in Generative AI
Laura Manduchi
Kushagra Pandey
Robert Bamler
Ryan Cotterell
Sina Daubener
...
F. Wenzel
Frank Wood
Stephan Mandt
Vincent Fortuin
Vincent Fortuin
54
17
0
28 Feb 2024
RoCoIns: Enhancing Robustness of Large Language Models through
  Code-Style Instructions
RoCoIns: Enhancing Robustness of Large Language Models through Code-Style Instructions
Yuan Zhang
Xiao Wang
Zhiheng Xi
Han Xia
Tao Gui
Qi Zhang
Xuanjing Huang
26
3
0
26 Feb 2024
Evaluating Robustness of Generative Search Engine on Adversarial Factual
  Questions
Evaluating Robustness of Generative Search Engine on Adversarial Factual Questions
Xuming Hu
Xiaochuan Li
Junzhe Chen
Yinghui Li
Yangning Li
...
Yasheng Wang
Qun Liu
Lijie Wen
Philip S. Yu
Zhijiang Guo
AAML
ELM
16
5
0
25 Feb 2024
Prompt Perturbation Consistency Learning for Robust Language Models
Prompt Perturbation Consistency Learning for Robust Language Models
Yao Qiang
Subhrangshu Nandi
Ninareh Mehrabi
Greg Ver Steeg
Anoop Kumar
Anna Rumshisky
Aram Galstyan
22
6
0
24 Feb 2024
Towards Robust Instruction Tuning on Multimodal Large Language Models
Towards Robust Instruction Tuning on Multimodal Large Language Models
Wei Han
Hui Chen
Soujanya Poria
MLLM
44
0
0
22 Feb 2024
Dynamic Evaluation of Large Language Models by Meta Probing Agents
Dynamic Evaluation of Large Language Models by Meta Probing Agents
Kaijie Zhu
Jindong Wang
Qinlin Zhao
Ruochen Xu
Xing Xie
35
30
0
21 Feb 2024
RITFIS: Robust input testing framework for LLMs-based intelligent
  software
RITFIS: Robust input testing framework for LLMs-based intelligent software
Ming-Ming Xiao
Yan Xiao
Hai Dong
Shunhui Ji
Pengcheng Zhang
AAML
33
5
0
21 Feb 2024
PhaseEvo: Towards Unified In-Context Prompt Optimization for Large
  Language Models
PhaseEvo: Towards Unified In-Context Prompt Optimization for Large Language Models
Wendi Cui
Jiaxin Zhang
Zhuohang Li
Hao Sun
Damien Lopez
Kamalika Das
Bradley Malin
Kumar Sricharan
14
6
0
17 Feb 2024
Contrastive Instruction Tuning
Contrastive Instruction Tuning
Tianyi Yan
Fei Wang
James Y. Huang
Wenxuan Zhou
Fan Yin
Aram Galstyan
Wenpeng Yin
Muhao Chen
ALM
5
5
0
17 Feb 2024
ToolSword: Unveiling Safety Issues of Large Language Models in Tool
  Learning Across Three Stages
ToolSword: Unveiling Safety Issues of Large Language Models in Tool Learning Across Three Stages
Junjie Ye
Sixian Li
Guanyu Li
Caishuang Huang
Songyang Gao
Yilong Wu
Qi Zhang
Tao Gui
Xuanjing Huang
LLMAG
22
16
0
16 Feb 2024
Large Language Models for the Automated Analysis of Optimization
  Algorithms
Large Language Models for the Automated Analysis of Optimization Algorithms
Camilo Chacón Sartori
Christian Blum
Gabriela Ochoa
19
2
0
13 Feb 2024
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Whispers in the Machine: Confidentiality in LLM-integrated Systems
Jonathan Evertz
Merlin Chlosta
Lea Schonherr
Thorsten Eisenhofer
69
15
0
10 Feb 2024
Measuring Implicit Bias in Explicitly Unbiased Large Language Models
Measuring Implicit Bias in Explicitly Unbiased Large Language Models
Xuechunzi Bai
Angelina Wang
Ilia Sucholutsky
Thomas L. Griffiths
94
27
0
06 Feb 2024
A Cross-Language Investigation into Jailbreak Attacks in Large Language
  Models
A Cross-Language Investigation into Jailbreak Attacks in Large Language Models
Jie Li
Yi Liu
Chongyang Liu
Ling Shi
Xiaoning Ren
Yaowen Zheng
Yang Liu
Yinxing Xue
AAML
9
23
0
30 Jan 2024
An Empirical Study on Large Language Models in Accuracy and Robustness
  under Chinese Industrial Scenarios
An Empirical Study on Large Language Models in Accuracy and Robustness under Chinese Industrial Scenarios
Zongjie Li
Wenying Qiu
Pingchuan Ma
Yichen Li
You Li
Sijia He
Baozheng Jiang
Shuai Wang
Weixi Gu
13
2
0
27 Jan 2024
Benchmarking Large Multimodal Models against Common Corruptions
Benchmarking Large Multimodal Models against Common Corruptions
Jiawei Zhang
Tianyu Pang
Chao Du
Yi Ren
Bo-wen Li
Min-Bin Lin
MLLM
22
14
0
22 Jan 2024
Deciphering Textual Authenticity: A Generalized Strategy through the
  Lens of Large Language Semantics for Detecting Human vs. Machine-Generated
  Text
Deciphering Textual Authenticity: A Generalized Strategy through the Lens of Large Language Semantics for Detecting Human vs. Machine-Generated Text
Mazal Bethany
Brandon Wherry
Emet Bethany
Nishant Vishwamitra
Anthony Rios
Peyman Najafirad
DeLMO
15
3
0
17 Jan 2024
Risk Taxonomy, Mitigation, and Assessment Benchmarks of Large Language
  Model Systems
Risk Taxonomy, Mitigation, and Assessment Benchmarks of Large Language Model Systems
Tianyu Cui
Yanling Wang
Chuanpu Fu
Yong Xiao
Sijia Li
...
Junwu Xiong
Xinyu Kong
Zujie Wen
Ke Xu
Qi Li
52
56
0
11 Jan 2024
METAL: Metamorphic Testing Framework for Analyzing Large-Language Model
  Qualities
METAL: Metamorphic Testing Framework for Analyzing Large-Language Model Qualities
Sangwon Hyun
Mingyu Guo
Muhammad Ali Babar
17
8
0
11 Dec 2023
A Survey on Large Language Model (LLM) Security and Privacy: The Good,
  the Bad, and the Ugly
A Survey on Large Language Model (LLM) Security and Privacy: The Good, the Bad, and the Ugly
Yifan Yao
Jinhao Duan
Kaidi Xu
Yuanfang Cai
Eric Sun
Yue Zhang
PILM
ELM
24
463
0
04 Dec 2023
Unnatural Error Correction: GPT-4 Can Almost Perfectly Handle Unnatural
  Scrambled Text
Unnatural Error Correction: GPT-4 Can Almost Perfectly Handle Unnatural Scrambled Text
Qi Cao
Takeshi Kojima
Yutaka Matsuo
Yusuke Iwasawa
10
18
0
30 Nov 2023
Walking a Tightrope -- Evaluating Large Language Models in High-Risk
  Domains
Walking a Tightrope -- Evaluating Large Language Models in High-Risk Domains
Chia-Chien Hung
Wiem Ben-Rim
Lindsay Frost
Lars Bruckner
Carolin (Haas) Lawrence
AILaw
ALM
ELM
17
9
0
25 Nov 2023
Hijacking Large Language Models via Adversarial In-Context Learning
Hijacking Large Language Models via Adversarial In-Context Learning
Yao Qiang
Xiangyu Zhou
Dongxiao Zhu
30
32
0
16 Nov 2023
Automatic Engineering of Long Prompts
Automatic Engineering of Long Prompts
Cho-Jui Hsieh
Si Si
Felix X. Yu
Inderjit S. Dhillon
VLM
11
8
0
16 Nov 2023
How Trustworthy are Open-Source LLMs? An Assessment under Malicious
  Demonstrations Shows their Vulnerabilities
How Trustworthy are Open-Source LLMs? An Assessment under Malicious Demonstrations Shows their Vulnerabilities
Lingbo Mo
Boshi Wang
Muhao Chen
Huan Sun
14
25
0
15 Nov 2023
A Wolf in Sheep's Clothing: Generalized Nested Jailbreak Prompts can
  Fool Large Language Models Easily
A Wolf in Sheep's Clothing: Generalized Nested Jailbreak Prompts can Fool Large Language Models Easily
Peng Ding
Jun Kuang
Dan Ma
Xuezhi Cao
Yunsen Xian
Jiajun Chen
Shujian Huang
AAML
12
94
0
14 Nov 2023
Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of
  LLMs through a Global Scale Prompt Hacking Competition
Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs through a Global Scale Prompt Hacking Competition
Sander Schulhoff
Jeremy Pinto
Anaum Khan
Louis-Franccois Bouchard
Chenglei Si
Svetlina Anati
Valen Tagliabue
Anson Liu Kost
Christopher Carnahan
Jordan L. Boyd-Graber
SILM
16
41
0
24 Oct 2023
AutoDAN: Interpretable Gradient-Based Adversarial Attacks on Large
  Language Models
AutoDAN: Interpretable Gradient-Based Adversarial Attacks on Large Language Models
Sicheng Zhu
Ruiyi Zhang
Bang An
Gang Wu
Joe Barrow
Zichao Wang
Furong Huang
A. Nenkova
Tong Sun
SILM
AAML
25
40
0
23 Oct 2023
She had Cobalt Blue Eyes: Prompt Testing to Create Aligned and
  Sustainable Language Models
She had Cobalt Blue Eyes: Prompt Testing to Create Aligned and Sustainable Language Models
Veronica Chatrath
Oluwanifemi Bamgbose
Shaina Raza
ALM
ELM
11
0
0
20 Oct 2023
An LLM can Fool Itself: A Prompt-Based Adversarial Attack
An LLM can Fool Itself: A Prompt-Based Adversarial Attack
Xilie Xu
Keyi Kong
Ning Liu
Li-zhen Cui
Di Wang
Jingfeng Zhang
Mohan S. Kankanhalli
AAML
SILM
17
68
0
20 Oct 2023
Assessing the Reliability of Large Language Model Knowledge
Assessing the Reliability of Large Language Model Knowledge
Weixuan Wang
Barry Haddow
Alexandra Birch
Wei Peng
KELM
HILM
60
14
0
15 Oct 2023
The Emergence of Reproducibility and Generalizability in Diffusion
  Models
The Emergence of Reproducibility and Generalizability in Diffusion Models
Huijie Zhang
Jinfan Zhou
Yifu Lu
Minzhe Guo
Peng Wang
Liyue Shen
Qing Qu
DiffM
18
2
0
08 Oct 2023
Fool Your (Vision and) Language Model With Embarrassingly Simple
  Permutations
Fool Your (Vision and) Language Model With Embarrassingly Simple Permutations
Yongshuo Zong
Tingyang Yu
Ruchika Chavhan
Bingchen Zhao
Timothy M. Hospedales
MLLM
AAML
LRM
18
17
0
02 Oct 2023
Meta Semantic Template for Evaluation of Large Language Models
Meta Semantic Template for Evaluation of Large Language Models
Yachuan Liu
Liang Chen
Jindong Wang
Qiaozhu Mei
Xing Xie
15
0
0
01 Oct 2023
Intuitive or Dependent? Investigating LLMs' Behavior Style to
  Conflicting Prompts
Intuitive or Dependent? Investigating LLMs' Behavior Style to Conflicting Prompts
Jiahao Ying
Yixin Cao
Kai Xiong
Yidong He
Long Cui
Yongbin Liu
23
7
0
29 Sep 2023
DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks
DyVal: Dynamic Evaluation of Large Language Models for Reasoning Tasks
A. Maritan
Jiaao Chen
S. Dey
Luca Schenato
Diyi Yang
Xing Xie
ELM
LRM
14
42
0
29 Sep 2023
HANS, are you clever? Clever Hans Effect Analysis of Neural Systems
Leonardo Ranaldi
Fabio Massimo Zanzotto
18
1
0
21 Sep 2023
Foundation Metrics for Evaluating Effectiveness of Healthcare
  Conversations Powered by Generative AI
Foundation Metrics for Evaluating Effectiveness of Healthcare Conversations Powered by Generative AI
Mahyar Abbasian
Elahe Khatibi
Iman Azimi
David Oniani
Zahra Shakeri Hossein Abad
...
Bryant Lin
Olivier Gevaert
Li-Jia Li
Ramesh C. Jain
Amir M. Rahmani
LM&MA
ELM
AI4MH
20
63
0
21 Sep 2023
How Robust is Google's Bard to Adversarial Image Attacks?
How Robust is Google's Bard to Adversarial Image Attacks?
Yinpeng Dong
Huanran Chen
Jiawei Chen
Zhengwei Fang
X. Yang
Yichi Zhang
Yu Tian
Hang Su
Jun Zhu
AAML
8
100
0
21 Sep 2023
Beyond Static Datasets: A Deep Interaction Approach to LLM Evaluation
Beyond Static Datasets: A Deep Interaction Approach to LLM Evaluation
Jiatong Li
Rui Li
Qi Liu
19
14
0
08 Sep 2023
MathAttack: Attacking Large Language Models Towards Math Solving Ability
MathAttack: Attacking Large Language Models Towards Math Solving Ability
Zihao Zhou
Qiufeng Wang
Mingyu Jin
Jie Yao
Jianan Ye
Wei Liu
Wei Wang
Xiaowei Huang
Kaizhu Huang
AAML
16
22
0
04 Sep 2023
Previous
123
Next