ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2203.01677
  4. Cited By
Detection of Word Adversarial Examples in Text Classification: Benchmark
  and Baseline via Robust Density Estimation

Detection of Word Adversarial Examples in Text Classification: Benchmark and Baseline via Robust Density Estimation

3 March 2022
Kiyoon Yoo
Jangho Kim
Jiho Jang
Nojun Kwak
ArXivPDFHTML

Papers citing "Detection of Word Adversarial Examples in Text Classification: Benchmark and Baseline via Robust Density Estimation"

27 / 27 papers shown
Title
Uncertainty-aware abstention in medical diagnosis based on medical texts
Uncertainty-aware abstention in medical diagnosis based on medical texts
Artem Vazhentsev
Ivan Sviridov
Alvard Barseghyan
Gleb Kuzmin
Alexander Panchenko
A. Nesterov
Artem Shelmanov
Maxim Panov
53
0
0
25 Feb 2025
Can Your Uncertainty Scores Detect Hallucinated Entity?
Can Your Uncertainty Scores Detect Hallucinated Entity?
Min-Hsuan Yeh
Max Kamachee
Seongheon Park
Yixuan Li
HILM
49
1
0
17 Feb 2025
A Stitch in Time Saves Nine: Small VLM is a Precise Guidance for
  Accelerating Large VLMs
A Stitch in Time Saves Nine: Small VLM is a Precise Guidance for Accelerating Large VLMs
Wangbo Zhao
Yizeng Han
Jiasheng Tang
Z. Li
Yibing Song
K. Wang
Zhangyang Wang
Yang You
83
7
0
04 Dec 2024
Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph
Benchmarking Uncertainty Quantification Methods for Large Language Models with LM-Polygraph
Roman Vashurin
Ekaterina Fadeeva
Artem Vazhentsev
Akim Tsvigun
Daniil Vasilev
...
Timothy Baldwin
Timothy Baldwin
Maxim Panov
Artem Shelmanov
Artem Shelmanov
HILM
66
8
0
21 Jun 2024
Adversarial Attacks and Dimensionality in Text Classifiers
Adversarial Attacks and Dimensionality in Text Classifiers
Nandish Chattopadhyay
Atreya Goswami
Anupam Chattopadhyay
SILM
AAML
19
1
0
03 Apr 2024
API Is Enough: Conformal Prediction for Large Language Models Without
  Logit-Access
API Is Enough: Conformal Prediction for Large Language Models Without Logit-Access
Jiayuan Su
Jing Luo
Hongwei Wang
Lu Cheng
74
16
0
02 Mar 2024
Calibrating Large Language Models with Sample Consistency
Calibrating Large Language Models with Sample Consistency
Qing Lyu
Kumar Shridhar
Chaitanya Malaviya
Li Zhang
Yanai Elazar
Niket Tandon
Marianna Apidianaki
Mrinmaya Sachan
Chris Callison-Burch
43
23
0
21 Feb 2024
Efficient Non-Parametric Uncertainty Quantification for Black-Box Large
  Language Models and Decision Planning
Efficient Non-Parametric Uncertainty Quantification for Black-Box Large Language Models and Decision Planning
Yao-Hung Tsai
Walter Talbott
Jian Zhang
LLMAG
16
4
0
01 Feb 2024
A Survey of Confidence Estimation and Calibration in Large Language
  Models
A Survey of Confidence Estimation and Calibration in Large Language Models
Jiahui Geng
Fengyu Cai
Yuxia Wang
Heinz Koeppl
Preslav Nakov
Iryna Gurevych
UQCV
41
54
0
14 Nov 2023
Efficient Black-Box Adversarial Attacks on Neural Text Detectors
Efficient Black-Box Adversarial Attacks on Neural Text Detectors
Vitalii Fishchuk
Daniel Braun
AAML
DeLMO
13
4
0
03 Nov 2023
Toward Stronger Textual Attack Detectors
Toward Stronger Textual Attack Detectors
Pierre Colombo
Marine Picot
Nathan Noiry
Guillaume Staerman
Pablo Piantanida
44
5
0
21 Oct 2023
Certifying LLM Safety against Adversarial Prompting
Certifying LLM Safety against Adversarial Prompting
Aounon Kumar
Chirag Agarwal
Suraj Srinivas
Aaron Jiaxun Li
S. Feizi
Himabindu Lakkaraju
AAML
27
164
0
06 Sep 2023
Text-CRS: A Generalized Certified Robustness Framework against Textual
  Adversarial Attacks
Text-CRS: A Generalized Certified Robustness Framework against Textual Adversarial Attacks
Xinyu Zhang
Hanbin Hong
Yuan Hong
Peng Huang
Binghui Wang
Zhongjie Ba
Kui Ren
SILM
29
18
0
31 Jul 2023
Interpretability and Transparency-Driven Detection and Transformation of
  Textual Adversarial Examples (IT-DT)
Interpretability and Transparency-Driven Detection and Transformation of Textual Adversarial Examples (IT-DT)
Bushra Sabir
Muhammad Ali Babar
Sharif Abuadbba
SILM
34
8
0
03 Jul 2023
On the Universal Adversarial Perturbations for Efficient Data-free
  Adversarial Detection
On the Universal Adversarial Perturbations for Efficient Data-free Adversarial Detection
Songyang Gao
Shihan Dou
Qi Zhang
Xuanjing Huang
Jin Ma
Yingchun Shan
AAML
9
3
0
27 Jun 2023
VoteTRANS: Detecting Adversarial Text without Training by Voting on Hard
  Labels of Transformations
VoteTRANS: Detecting Adversarial Text without Training by Voting on Hard Labels of Transformations
Hoang-Quoc Nguyen-Son
Seira Hidano
Kazuhide Fukushima
S. Kiyomoto
Isao Echizen
26
0
0
02 Jun 2023
Query-Efficient Black-Box Red Teaming via Bayesian Optimization
Query-Efficient Black-Box Red Teaming via Bayesian Optimization
Deokjae Lee
JunYeong Lee
Jung-Woo Ha
Jin-Hwa Kim
Sang-Woo Lee
Hwaran Lee
Hyun Oh Song
AAML
21
23
0
27 May 2023
Can Large Language Models Be an Alternative to Human Evaluations?
Can Large Language Models Be an Alternative to Human Evaluations?
Cheng-Han Chiang
Hung-yi Lee
ALM
LM&MA
224
571
0
03 May 2023
Verifying the Robustness of Automatic Credibility Assessment
Verifying the Robustness of Automatic Credibility Assessment
Piotr Przybyła
A. Shvets
Horacio Saggion
DeLMO
AAML
30
6
0
14 Mar 2023
ADDMU: Detection of Far-Boundary Adversarial Examples with Data and
  Model Uncertainty Estimation
ADDMU: Detection of Far-Boundary Adversarial Examples with Data and Model Uncertainty Estimation
Fan Yin
Yao Li
Cho-Jui Hsieh
Kai-Wei Chang
AAML
60
4
0
22 Oct 2022
Are Synonym Substitution Attacks Really Synonym Substitution Attacks?
Are Synonym Substitution Attacks Really Synonym Substitution Attacks?
Cheng-Han Chiang
Hunghuei Lee
AAML
33
5
0
06 Oct 2022
On deceiving malware classification with section injection
On deceiving malware classification with section injection
Adeilson Antonio da Silva
Maurício Pamplona Segundo
28
4
0
12 Aug 2022
Understanding, Detecting, and Separating Out-of-Distribution Samples and
  Adversarial Samples in Text Classification
Understanding, Detecting, and Separating Out-of-Distribution Samples and Adversarial Samples in Text Classification
Cheng-Han Chiang
Hung-yi Lee
OODD
23
1
0
09 Apr 2022
A Survey of Adversarial Defences and Robustness in NLP
A Survey of Adversarial Defences and Robustness in NLP
Shreyansh Goyal
Sumanth Doddapaneni
Mitesh M.Khapra
B. Ravindran
AAML
29
30
0
12 Mar 2022
Adversarial Examples Detection with Bayesian Neural Network
Adversarial Examples Detection with Bayesian Neural Network
Yao Li
Tongyi Tang
Cho-Jui Hsieh
T. C. Lee
GAN
AAML
30
3
0
18 May 2021
Robust Encodings: A Framework for Combating Adversarial Typos
Robust Encodings: A Framework for Combating Adversarial Typos
Erik Jones
Robin Jia
Aditi Raghunathan
Percy Liang
AAML
127
102
0
04 May 2020
Convolutional Neural Networks for Sentence Classification
Convolutional Neural Networks for Sentence Classification
Yoon Kim
AILaw
VLM
255
13,364
0
25 Aug 2014
1