ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2003.12613
  4. Cited By
Adaptive Reward-Poisoning Attacks against Reinforcement Learning
v1v2 (latest)

Adaptive Reward-Poisoning Attacks against Reinforcement Learning

International Conference on Machine Learning (ICML), 2020
27 March 2020
Xuezhou Zhang
Yuzhe Ma
Adish Singla
Xiaojin Zhu
    AAML
ArXiv (abs)PDFHTML

Papers citing "Adaptive Reward-Poisoning Attacks against Reinforcement Learning"

50 / 89 papers shown
Exposing Vulnerabilities in RL: A Novel Stealthy Backdoor Attack through Reward Poisoning
Exposing Vulnerabilities in RL: A Novel Stealthy Backdoor Attack through Reward Poisoning
Bokang Zhang
Chaojun Lu
Jianhui Li
Junfeng Wu
AAML
181
0
0
27 Nov 2025
Diffusion Guided Adversarial State Perturbations in Reinforcement Learning
Diffusion Guided Adversarial State Perturbations in Reinforcement Learning
Xiaolin Sun
Feidi Liu
Zhengming Ding
Zizhan Zheng
AAML
193
0
0
10 Nov 2025
Agentic AI Security: Threats, Defenses, Evaluation, and Open Challenges
Agentic AI Security: Threats, Defenses, Evaluation, and Open Challenges
Shrestha Datta
Shahriar Kabir Nahin
Anshuman Chhabra
P. Mohapatra
LLMAGLM&Ro
414
8
0
27 Oct 2025
Provably Invincible Adversarial Attacks on Reinforcement Learning Systems: A Rate-Distortion Information-Theoretic Approach
Provably Invincible Adversarial Attacks on Reinforcement Learning Systems: A Rate-Distortion Information-Theoretic Approach
Ziqing Lu
Lifeng Lai
Weiyu Xu
AAML
118
0
0
15 Oct 2025
Density-Ratio Weighted Behavioral Cloning: Learning Control Policies from Corrupted Datasets
Density-Ratio Weighted Behavioral Cloning: Learning Control Policies from Corrupted Datasets
Shriram Karpoora Sundara Pandian
Ali Baheri
OffRL
215
0
0
01 Oct 2025
Constrained Black-Box Attacks Against Cooperative Multi-Agent Reinforcement Learning
Constrained Black-Box Attacks Against Cooperative Multi-Agent Reinforcement Learning
Amine Andam
Jamal Bentahar
Mustapha Hedabou
AAML
130
0
0
12 Aug 2025
Policy Disruption in Reinforcement Learning:Adversarial Attack with Large Language Models and Critical State Identification
Policy Disruption in Reinforcement Learning:Adversarial Attack with Large Language Models and Critical State Identification
Junyong Jiang
Buwei Tian
Chenxing Xu
Songze Li
Lu Dong
AAML
164
1
0
24 Jul 2025
Collapsing Sequence-Level Data-Policy Coverage via Poisoning Attack in Offline Reinforcement Learning
Collapsing Sequence-Level Data-Policy Coverage via Poisoning Attack in Offline Reinforcement LearningConference on Uncertainty in Artificial Intelligence (UAI), 2025
Xue Zhou
Dapeng Man
Chen Xu
Fanyi Zeng
Tao Liu
Huan Wang
Shucheng He
Chaoyang Gao
Wu Yang
OffRL
217
0
0
12 Jun 2025
Can In-Context Reinforcement Learning Recover From Reward Poisoning Attacks?
Can In-Context Reinforcement Learning Recover From Reward Poisoning Attacks?
Paulius Sasnauskas
Yiğit Yalın
Goran Radanović
265
0
0
07 Jun 2025
Optimally Installing Strict Equilibria
Jeremy McMahan
Young Wu
Yudong Chen
Xiaojin Zhu
Qiaomin Xie
339
0
0
05 Mar 2025
Provably Robust Federated Reinforcement Learning
Provably Robust Federated Reinforcement LearningThe Web Conference (WWW), 2025
Minghong Fang
Xilong Wang
Neil Zhenqiang Gong
FedML
329
12
0
12 Feb 2025
Reinforcement Teaching
Reinforcement Teaching
Alex Lewandowski
Calarina Muslimani
Dale Schuurmans
Matthew E. Taylor
Jun Luo
456
2
0
28 Jan 2025
Position: Mind the Gap-the Growing Disconnect Between Established Vulnerability Disclosure and AI Security
Position: Mind the Gap-the Growing Disconnect Between Established Vulnerability Disclosure and AI Security
L. Bieringer
Sean McGregor
Nicole Nichols
Kevin Paeth
Jochen Stängler
Andreas Wespi
Alexandre Alahi
Kathrin Grosse
532
0
0
19 Dec 2024
Online Poisoning Attack Against Reinforcement Learning under Black-box
  Environments
Online Poisoning Attack Against Reinforcement Learning under Black-box Environments
Jianhui Li
Bokang Zhang
Junfeng Wu
AAMLOffRLOnRL
345
4
0
01 Dec 2024
Provably Efficient Action-Manipulation Attack Against Continuous
  Reinforcement Learning
Provably Efficient Action-Manipulation Attack Against Continuous Reinforcement Learning
Zhi Luo
Xiaoyu Yang
Pan Zhou
D. Wang
AAML
266
1
0
20 Nov 2024
SpecGuard: Specification Aware Recovery for Robotic Autonomous Vehicles
  from Physical Attacks
SpecGuard: Specification Aware Recovery for Robotic Autonomous Vehicles from Physical AttacksConference on Computer and Communications Security (CCS), 2024
Pritam Dash
Ethan Chan
Karthik Pattabiraman
AAML
213
11
0
27 Aug 2024
CuDA2: An approach for Incorporating Traitor Agents into Cooperative
  Multi-Agent Systems
CuDA2: An approach for Incorporating Traitor Agents into Cooperative Multi-Agent Systems
Zhen Chen
Yong Liao
Youpeng Zhao
Zipeng Dai
Jian Zhao
AAML
210
1
0
25 Jun 2024
Defending Against Sophisticated Poisoning Attacks with RL-based
  Aggregation in Federated Learning
Defending Against Sophisticated Poisoning Attacks with RL-based Aggregation in Federated Learning
Yujing Wang
Hainan Zhang
Sijia Wen
Wangjie Qiu
Binghui Guo
AAML
217
3
0
20 Jun 2024
Exploring Fairness in Educational Data Mining in the Context of the
  Right to be Forgotten
Exploring Fairness in Educational Data Mining in the Context of the Right to be Forgotten
Wei Qian
Aobo Chen
Chenxu Zhao
Yangyi Li
Mengdi Huai
MU
377
2
0
27 May 2024
Policy Iteration for Two-Player General-Sum Stochastic Stackelberg Games
Policy Iteration for Two-Player General-Sum Stochastic Stackelberg Games
Mikoto Kudo
Yohei Akimoto
219
0
0
07 May 2024
Data Poisoning Attacks on Off-Policy Policy Evaluation Methods
Data Poisoning Attacks on Off-Policy Policy Evaluation Methods
Elita Lobo
Harvineet Singh
Marek Petrik
Cynthia Rudin
Himabindu Lakkaraju
250
3
0
06 Apr 2024
Strategizing against Q-learners: A Control-theoretical Approach
Strategizing against Q-learners: A Control-theoretical ApproachIEEE Control Systems Letters (L-CSS), 2024
Yuksel Arslantas
Ege Yuceel
Muhammed O. Sayin
326
6
0
13 Mar 2024
Belief-Enriched Pessimistic Q-Learning against Adversarial State
  Perturbations
Belief-Enriched Pessimistic Q-Learning against Adversarial State Perturbations
Xiaolin Sun
Zizhan Zheng
OnRL
295
3
0
06 Mar 2024
Stealthy Adversarial Attacks on Stochastic Multi-Armed Bandits
Stealthy Adversarial Attacks on Stochastic Multi-Armed Bandits
Zhiwei Wang
Huazheng Wang
Hongning Wang
AAML
320
1
0
21 Feb 2024
Beyond Worst-case Attacks: Robust RL with Adaptive Defense via
  Non-dominated Policies
Beyond Worst-case Attacks: Robust RL with Adaptive Defense via Non-dominated Policies
Xiangyu Liu
Chenghao Deng
Yanchao Sun
Yongyuan Liang
Furong Huang
AAML
365
10
0
20 Feb 2024
SUB-PLAY: Adversarial Policies against Partially Observed Multi-Agent
  Reinforcement Learning Systems
SUB-PLAY: Adversarial Policies against Partially Observed Multi-Agent Reinforcement Learning SystemsConference on Computer and Communications Security (CCS), 2024
Oubo Ma
Yuwen Pu
L. Du
Yang Dai
Ruo Wang
Xiaolei Liu
Yingcai Wu
Shouling Ji
AAML
302
14
0
06 Feb 2024
Assessing the Impact of Distribution Shift on Reinforcement Learning
  Performance
Assessing the Impact of Distribution Shift on Reinforcement Learning Performance
Ted Fujimoto
Joshua Suetterlein
Samrat Chatterjee
A. Ganguly
OffRL
264
9
0
05 Feb 2024
Camouflage Adversarial Attacks on Multiple Agent Systems
Camouflage Adversarial Attacks on Multiple Agent Systems
Ziqing Lu
Guanlin Liu
Lifeng Lai
Weiyu Xu
AAML
224
4
0
30 Jan 2024
Adaptive Discounting of Training Time Attacks
Adaptive Discounting of Training Time Attacks
Ridhima Bector
Abhay M. S. Aradhya
Chai Quek
Zinovi Rabinovich
AAML
257
0
0
05 Jan 2024
PGN: A perturbation generation network against deep reinforcement
  learning
PGN: A perturbation generation network against deep reinforcement learning
Xiangjuan Li
Feifan Li
Yang Li
Quanbiao Pan
AAML
129
2
0
20 Dec 2023
BadRL: Sparse Targeted Backdoor Attack Against Reinforcement Learning
BadRL: Sparse Targeted Backdoor Attack Against Reinforcement Learning
Jing Cui
Yufei Han
Yuzhe Ma
Jianbin Jiao
Junge Zhang
AAML
268
28
0
19 Dec 2023
Optimal Attack and Defense for Reinforcement Learning
Optimal Attack and Defense for Reinforcement LearningAAAI Conference on Artificial Intelligence (AAAI), 2023
Jeremy McMahan
Young Wu
Xiaojin Zhu
Qiaomin Xie
AAMLOffRL
328
20
0
30 Nov 2023
RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with
  Human Feedback in Large Language Models
RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with Human Feedback in Large Language Models
Zhenghao Hu
Junlin Wu
Muhao Chen
Yevgeniy Vorobeychik
Chaowei Xiao
AAML
261
32
0
16 Nov 2023
Optimal Cost Constrained Adversarial Attacks For Multiple Agent Systems
Optimal Cost Constrained Adversarial Attacks For Multiple Agent SystemsAnnual Conference on Information Sciences and Systems (CISS), 2023
Ziqing Lu
Guanlin Liu
Lifeng Lai
Weiyu Xu
AAML
213
3
0
01 Nov 2023
Minimally Modifying a Markov Game to Achieve Any Nash Equilibrium and
  Value
Minimally Modifying a Markov Game to Achieve Any Nash Equilibrium and ValueInternational Conference on Machine Learning (ICML), 2023
Young Wu
Jeremy McMahan
Yiding Chen
Yudong Chen
Xiaojin Zhu
Qiaomin Xie
508
3
0
01 Nov 2023
Corruption-Robust Offline Reinforcement Learning with General Function
  Approximation
Corruption-Robust Offline Reinforcement Learning with General Function ApproximationNeural Information Processing Systems (NeurIPS), 2023
Chen Ye
Rui Yang
Quanquan Gu
Tong Zhang
OffRL
463
30
0
23 Oct 2023
Adversarial Attacks on Combinatorial Multi-Armed Bandits
Adversarial Attacks on Combinatorial Multi-Armed BanditsInternational Conference on Machine Learning (ICML), 2023
R. Balasubramanian
Jiawei Li
Prasad Tadepalli
Huazheng Wang
Qingyun Wu
Haoyu Zhao
AAML
321
3
0
08 Oct 2023
Building Privacy-Preserving and Secure Geospatial Artificial
  Intelligence Foundation Models
Building Privacy-Preserving and Secure Geospatial Artificial Intelligence Foundation Models
Jinmeng Rao
Song Gao
Gengchen Mai
Joanna M. Wardlaw
378
29
0
29 Sep 2023
BRNES: Enabling Security and Privacy-aware Experience Sharing in
  Multiagent Robotic and Autonomous Systems
BRNES: Enabling Security and Privacy-aware Experience Sharing in Multiagent Robotic and Autonomous SystemsIEEE/RJS International Conference on Intelligent RObots and Systems (IROS), 2023
Md Tamjid Hossain
Hung M. La
S. Badsha
Anton Netchaev
165
4
0
02 Aug 2023
Efficient Adversarial Attacks on Online Multi-agent Reinforcement
  Learning
Efficient Adversarial Attacks on Online Multi-agent Reinforcement LearningNeural Information Processing Systems (NeurIPS), 2023
Guanlin Liu
Lifeng Lai
AAML
223
18
0
15 Jul 2023
Efficient Action Robust Reinforcement Learning with Probabilistic Policy
  Execution Uncertainty
Efficient Action Robust Reinforcement Learning with Probabilistic Policy Execution Uncertainty
Guanin Liu
Zhihan Zhou
Han Liu
Lifeng Lai
333
4
0
15 Jul 2023
Data Poisoning to Fake a Nash Equilibrium in Markov Games
Data Poisoning to Fake a Nash Equilibrium in Markov Games
Young Wu
Jeremy McMahan
Xiaojin Zhu
Qiaomin Xie
OffRL
319
2
0
13 Jun 2023
A Reminder of its Brittleness: Language Reward Shaping May Hinder
  Learning for Instruction Following Agents
A Reminder of its Brittleness: Language Reward Shaping May Hinder Learning for Instruction Following Agents
Sukai Huang
Nir Lipovetzky
Trevor Cohn
286
2
0
26 May 2023
Black-Box Targeted Reward Poisoning Attack Against Online Deep
  Reinforcement Learning
Black-Box Targeted Reward Poisoning Attack Against Online Deep Reinforcement Learning
Yinglun Xu
Gagandeep Singh
OffRLAAML
232
5
0
18 May 2023
Policy Resilience to Environment Poisoning Attacks on Reinforcement
  Learning
Policy Resilience to Environment Poisoning Attacks on Reinforcement Learning
Hang Xu
Xinghua Qu
Zinovi Rabinovich
256
3
0
24 Apr 2023
Local Environment Poisoning Attacks on Federated Reinforcement Learning
Local Environment Poisoning Attacks on Federated Reinforcement Learning
Evelyn Ma
Praneet Rathi
Rasoul Etesami
FedML
412
7
0
05 Mar 2023
Implicit Poisoning Attacks in Two-Agent Reinforcement Learning:
  Adversarial Policies for Training-Time Attacks
Implicit Poisoning Attacks in Two-Agent Reinforcement Learning: Adversarial Policies for Training-Time AttacksAdaptive Agents and Multi-Agent Systems (AAMAS), 2023
Mohammad Mohammadi
Jonathan Nöther
Debmalya Mandal
Adish Singla
Goran Radanović
AAMLOffRL
219
12
0
27 Feb 2023
Adversarial Attacks on Adversarial Bandits
Adversarial Attacks on Adversarial BanditsInternational Conference on Learning Representations (ICLR), 2023
Yuzhe Ma
Zhijin Zhou
AAML
184
10
0
30 Jan 2023
New Challenges in Reinforcement Learning: A Survey of Security and
  Privacy
New Challenges in Reinforcement Learning: A Survey of Security and PrivacyArtificial Intelligence Review (Artif Intell Rev), 2022
Yunjiao Lei
Dayong Ye
Sheng Shen
Yulei Sui
Tianqing Zhu
Wanlei Zhou
346
27
0
31 Dec 2022
Security of Deep Reinforcement Learning for Autonomous Driving: A Survey
Security of Deep Reinforcement Learning for Autonomous Driving: A Survey
Ambra Demontis
Srishti Gupta
Christian Scano
Luca Demetrio
Kathrin Grosse
Hsiao-Ying Lin
Chengfang Fang
Battista Biggio
Fabio Roli
AAML
387
4
0
12 Dec 2022
12
Next
Page 1 of 2