Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2505.15175
Cited By
A Linear Approach to Data Poisoning
21 May 2025
Diego Granziol
Donald Flynn
AAML
Re-assign community
ArXiv
PDF
HTML
Papers citing
"A Linear Approach to Data Poisoning"
17 / 17 papers shown
Title
Shadowcast: Stealthy Data Poisoning Attacks Against Vision-Language Models
Yuancheng Xu
Jiarui Yao
Manli Shu
Yanchao Sun
Zichu Wu
Ning Yu
Tom Goldstein
Furong Huang
AAML
76
18
0
05 Feb 2024
Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training
Evan Hubinger
Carson E. Denison
Jesse Mu
Mike Lambert
Meg Tong
...
Sören Mindermann
Ryan Greenblatt
Buck Shlegeris
Nicholas Schiefer
Ethan Perez
LLMAG
29
159
0
10 Jan 2024
The Stronger the Diffusion Model, the Easier the Backdoor: Data Poisoning to Induce Copyright Breaches Without Adjusting Finetuning Pipeline
Haonan Wang
Qianli Shen
Yao Tong
Yang Zhang
Kenji Kawaguchi
66
26
0
07 Jan 2024
From Trojan Horses to Castle Walls: Unveiling Bilateral Data Poisoning Effects in Diffusion Models
Zhuoshi Pan
Yuguang Yao
Gaowen Liu
Bingquan Shen
H. Vicky Zhao
Ramana Rao Kompella
Sijia Liu
DiffM
AAML
71
4
0
04 Nov 2023
Universal and Transferable Adversarial Attacks on Aligned Language Models
Andy Zou
Zifan Wang
Nicholas Carlini
Milad Nasr
J. Zico Kolter
Matt Fredrikson
152
1,376
0
27 Jul 2023
LLaMA: Open and Efficient Foundation Language Models
Hugo Touvron
Thibaut Lavril
Gautier Izacard
Xavier Martinet
Marie-Anne Lachaux
...
Faisal Azhar
Aurelien Rodriguez
Armand Joulin
Edouard Grave
Guillaume Lample
ALM
PILM
646
12,840
0
27 Feb 2023
BadGPT: Exploring Security Vulnerabilities of ChatGPT via Backdoor Attacks to InstructGPT
Jiawen Shi
Yixin Liu
Pan Zhou
Lichao Sun
SILM
16
77
0
21 Feb 2023
How to Backdoor Diffusion Models?
Sheng-Yen Chou
Pin-Yu Chen
Tsung-Yi Ho
DiffM
SILM
69
99
0
11 Dec 2022
BadPrompt: Backdoor Attacks on Continuous Prompts
Xiangrui Cai
Haidong Xu
Sihan Xu
Ying Zhang
Xiaojie Yuan
SILM
44
63
0
27 Nov 2022
Indiscriminate Poisoning Attacks on Unsupervised Contrastive Learning
Hao He
Kaiwen Zha
Dina Katabi
AAML
67
33
0
22 Feb 2022
Poisoning and Backdooring Contrastive Learning
Nicholas Carlini
Andreas Terzis
51
161
0
17 Jun 2021
Handcrafted Backdoors in Deep Neural Networks
Sanghyun Hong
Nicholas Carlini
Alexey Kurakin
74
74
0
08 Jun 2021
Learning Transferable Visual Models From Natural Language Supervision
Alec Radford
Jong Wook Kim
Chris Hallacy
Aditya A. Ramesh
Gabriel Goh
...
Amanda Askell
Pamela Mishkin
Jack Clark
Gretchen Krueger
Ilya Sutskever
CLIP
VLM
666
28,659
0
26 Feb 2021
Poisoned classifiers are not only backdoored, they are fundamentally broken
Mingjie Sun
Siddhant Agarwal
J. Zico Kolter
22
26
0
18 Oct 2020
PyTorch: An Imperative Style, High-Performance Deep Learning Library
Adam Paszke
Sam Gross
Francisco Massa
Adam Lerer
James Bradbury
...
Sasank Chilamkurthy
Benoit Steiner
Lu Fang
Junjie Bai
Soumith Chintala
ODL
224
42,038
0
03 Dec 2019
Spectral Signatures in Backdoor Attacks
Brandon Tran
Jerry Li
Aleksander Madry
AAML
65
778
0
01 Nov 2018
BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain
Tianyu Gu
Brendan Dolan-Gavitt
S. Garg
SILM
72
1,754
0
22 Aug 2017
1