ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2101.00288
  4. Cited By
Polyjuice: Generating Counterfactuals for Explaining, Evaluating, and
  Improving Models

Polyjuice: Generating Counterfactuals for Explaining, Evaluating, and Improving Models

1 January 2021
Tongshuang Wu
Marco Tulio Ribeiro
Jeffrey Heer
Daniel S. Weld
ArXivPDFHTML

Papers citing "Polyjuice: Generating Counterfactuals for Explaining, Evaluating, and Improving Models"

50 / 178 papers shown
Title
Towards a Non-Ideal Methodological Framework for Responsible ML
Towards a Non-Ideal Methodological Framework for Responsible ML
Ramaravind Kommiya Mothilal
Shion Guha
Syed Ishtiaque Ahmed
40
7
0
20 Jan 2024
An Empirical Study of Counterfactual Visualization to Support Visual
  Causal Inference
An Empirical Study of Counterfactual Visualization to Support Visual Causal Inference
Arran Zeyu Wang
D. Borland
David Gotz
CML
17
11
0
16 Jan 2024
Are self-explanations from Large Language Models faithful?
Are self-explanations from Large Language Models faithful?
Andreas Madsen
Sarath Chandar
Siva Reddy
LRM
30
24
0
15 Jan 2024
Sparsity-Guided Holistic Explanation for LLMs with Interpretable
  Inference-Time Intervention
Sparsity-Guided Holistic Explanation for LLMs with Interpretable Inference-Time Intervention
Zhen Tan
Tianlong Chen
Zhenyu (Allen) Zhang
Huan Liu
42
14
0
22 Dec 2023
InstructPipe: Generating Visual Blocks Pipelines with Human Instructions and LLMs
InstructPipe: Generating Visual Blocks Pipelines with Human Instructions and LLMs
Zhongyi Zhou
Jing Jin
Vrushank Phadnis
Xiuxiu Yuan
Jun Jiang
...
A. Olwal
David Kim
Ram Iyengar
Na Li
Ruofei Du
28
5
0
15 Dec 2023
Using Captum to Explain Generative Language Models
Using Captum to Explain Generative Language Models
Vivek Miglani
Aobo Yang
Aram H. Markosyan
Diego Garcia-Olano
Narine Kokhlikyan
18
26
0
09 Dec 2023
TIBET: Identifying and Evaluating Biases in Text-to-Image Generative
  Models
TIBET: Identifying and Evaluating Biases in Text-to-Image Generative Models
Aditya Chinchure
Pushkar Shukla
Gaurav Bhatt
Kiri Salij
K. Hosanagar
Leonid Sigal
Matthew A. Turk
21
22
0
03 Dec 2023
SocialCounterfactuals: Probing and Mitigating Intersectional Social
  Biases in Vision-Language Models with Counterfactual Examples
SocialCounterfactuals: Probing and Mitigating Intersectional Social Biases in Vision-Language Models with Counterfactual Examples
Phillip Howard
Avinash Madasu
Tiep Le
Gustavo Lujan Moreno
Anahita Bhiwandiwalla
Vasudev Lal
45
16
0
30 Nov 2023
Attribution and Alignment: Effects of Local Context Repetition on
  Utterance Production and Comprehension in Dialogue
Attribution and Alignment: Effects of Local Context Repetition on Utterance Production and Comprehension in Dialogue
Aron Molnar
Jaap Jumelet
Mario Giulianelli
Arabella J. Sinclair
14
2
0
21 Nov 2023
Measuring and Improving Attentiveness to Partial Inputs with
  Counterfactuals
Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals
Yanai Elazar
Bhargavi Paranjape
Hao Peng
Sarah Wiegreffe
Khyathi Raghavi
Vivek Srikumar
Sameer Singh
Noah A. Smith
AAML
OOD
21
0
0
16 Nov 2023
Using Natural Language Explanations to Improve Robustness of In-context
  Learning
Using Natural Language Explanations to Improve Robustness of In-context Learning
Xuanli He
Yuxiang Wu
Oana-Maria Camburu
Pasquale Minervini
Pontus Stenetorp
AAML
23
1
0
13 Nov 2023
Interpreting Pretrained Language Models via Concept Bottlenecks
Interpreting Pretrained Language Models via Concept Bottlenecks
Zhen Tan
Lu Cheng
Song Wang
Yuan Bo
Jundong Li
Huan Liu
LRM
24
20
0
08 Nov 2023
Quantifying Uncertainty in Natural Language Explanations of Large
  Language Models
Quantifying Uncertainty in Natural Language Explanations of Large Language Models
Sree Harsha Tanneru
Chirag Agarwal
Himabindu Lakkaraju
LRM
25
13
0
06 Nov 2023
"Honey, Tell Me What's Wrong", Global Explanation of Textual
  Discriminative Models through Cooperative Generation
"Honey, Tell Me What's Wrong", Global Explanation of Textual Discriminative Models through Cooperative Generation
Antoine Chaffin
Julien Delaunay
8
0
0
27 Oct 2023
Break it, Imitate it, Fix it: Robustness by Generating Human-Like
  Attacks
Break it, Imitate it, Fix it: Robustness by Generating Human-Like Attacks
Aradhana Sinha
Ananth Balashankar
Ahmad Beirami
Thi Avrahami
Jilin Chen
Alex Beutel
AAML
27
4
0
25 Oct 2023
Sum-of-Parts: Self-Attributing Neural Networks with End-to-End Learning of Feature Groups
Sum-of-Parts: Self-Attributing Neural Networks with End-to-End Learning of Feature Groups
Weiqiu You
Helen Qu
Marco Gatti
Bhuvnesh Jain
Eric Wong
FAtt
FaML
32
4
0
25 Oct 2023
Towards Conceptualization of "Fair Explanation": Disparate Impacts of
  anti-Asian Hate Speech Explanations on Content Moderators
Towards Conceptualization of "Fair Explanation": Disparate Impacts of anti-Asian Hate Speech Explanations on Content Moderators
Tin Nguyen
Jiannan Xu
Aayushi Roy
Hal Daumé
Marine Carpuat
25
5
0
23 Oct 2023
EXPLAIN, EDIT, GENERATE: Rationale-Sensitive Counterfactual Data
  Augmentation for Multi-hop Fact Verification
EXPLAIN, EDIT, GENERATE: Rationale-Sensitive Counterfactual Data Augmentation for Multi-hop Fact Verification
Yingjie Zhu
Jiasheng Si
Yibo Zhao
Haiyang Zhu
Deyu Zhou
Yulan He
36
6
0
23 Oct 2023
Faithfulness Measurable Masked Language Models
Faithfulness Measurable Masked Language Models
Andreas Madsen
Siva Reddy
Sarath Chandar
38
3
0
11 Oct 2023
InterroLang: Exploring NLP Models and Datasets through Dialogue-based
  Explanations
InterroLang: Exploring NLP Models and Datasets through Dialogue-based Explanations
Nils Feldhus
Qianli Wang
Tatiana Anikina
Sahil Chopra
Cennet Oguz
Sebastian Möller
29
9
0
09 Oct 2023
Model Compression in Practice: Lessons Learned from Practitioners
  Creating On-device Machine Learning Experiences
Model Compression in Practice: Lessons Learned from Practitioners Creating On-device Machine Learning Experiences
Fred Hohman
Mary Beth Kery
Donghao Ren
Dominik Moritz
19
16
0
06 Oct 2023
From Language Modeling to Instruction Following: Understanding the
  Behavior Shift in LLMs after Instruction Tuning
From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning
Xuansheng Wu
Wenlin Yao
Jianshu Chen
Xiaoman Pan
Xiaoyang Wang
Ninghao Liu
Dong Yu
LRM
20
26
0
30 Sep 2023
EvalLM: Interactive Evaluation of Large Language Model Prompts on
  User-Defined Criteria
EvalLM: Interactive Evaluation of Large Language Model Prompts on User-Defined Criteria
Tae Soo Kim
Yoonjoo Lee
Jamin Shin
Young-Ho Kim
Juho Kim
21
40
0
24 Sep 2023
Towards LLM-guided Causal Explainability for Black-box Text Classifiers
Towards LLM-guided Causal Explainability for Black-box Text Classifiers
Amrita Bhattacharjee
Raha Moraffah
Joshua Garland
Huan Liu
11
33
0
23 Sep 2023
COCO-Counterfactuals: Automatically Constructed Counterfactual Examples
  for Image-Text Pairs
COCO-Counterfactuals: Automatically Constructed Counterfactual Examples for Image-Text Pairs
Tiep Le
Vasudev Lal
Phillip Howard
DiffM
26
21
0
23 Sep 2023
CATfOOD: Counterfactual Augmented Training for Improving Out-of-Domain
  Performance and Calibration
CATfOOD: Counterfactual Augmented Training for Improving Out-of-Domain Performance and Calibration
Rachneet Sachdeva
Martin Tutek
Iryna Gurevych
OODD
10
10
0
14 Sep 2023
Explainability for Large Language Models: A Survey
Explainability for Large Language Models: A Survey
Haiyan Zhao
Hanjie Chen
Fan Yang
Ninghao Liu
Huiqi Deng
Hengyi Cai
Shuaiqiang Wang
Dawei Yin
Mengnan Du
LRM
19
407
0
02 Sep 2023
PEANUT: A Human-AI Collaborative Tool for Annotating Audio-Visual Data
PEANUT: A Human-AI Collaborative Tool for Annotating Audio-Visual Data
Zheng Zhang
Zheng Ning
Chenliang Xu
Yapeng Tian
Toby Jia-Jun Li
59
6
0
27 Jul 2023
CommonsenseVIS: Visualizing and Understanding Commonsense Reasoning
  Capabilities of Natural Language Models
CommonsenseVIS: Visualizing and Understanding Commonsense Reasoning Capabilities of Natural Language Models
Xingbo Wang
Renfei Huang
Zhihua Jin
Tianqing Fang
Huamin Qu
VLM
ReLM
LRM
30
1
0
23 Jul 2023
Do Models Explain Themselves? Counterfactual Simulatability of Natural
  Language Explanations
Do Models Explain Themselves? Counterfactual Simulatability of Natural Language Explanations
Yanda Chen
Ruiqi Zhong
Narutatsu Ri
Chen Zhao
He He
Jacob Steinhardt
Zhou Yu
Kathleen McKeown
LRM
24
47
0
17 Jul 2023
Power-up! What Can Generative Models Do for Human Computation Workflows?
Power-up! What Can Generative Models Do for Human Computation Workflows?
Garrett Allen
Gaole He
U. Gadiraju
33
3
0
05 Jul 2023
Concept-Based Explanations to Test for False Causal Relationships
  Learned by Abusive Language Classifiers
Concept-Based Explanations to Test for False Causal Relationships Learned by Abusive Language Classifiers
I. Nejadgholi
S. Kiritchenko
Kathleen C. Fraser
Esma Balkir
21
0
0
04 Jul 2023
On Evaluating and Mitigating Gender Biases in Multilingual Settings
On Evaluating and Mitigating Gender Biases in Multilingual Settings
Aniket Vashishtha
Kabir Ahuja
Sunayana Sitaram
13
23
0
04 Jul 2023
Bring Your Own Data! Self-Supervised Evaluation for Large Language
  Models
Bring Your Own Data! Self-Supervised Evaluation for Large Language Models
Neel Jain
Khalid Saifullah
Yuxin Wen
John Kirchenbauer
Manli Shu
Aniruddha Saha
Micah Goldblum
Jonas Geiping
Tom Goldstein
ALM
ELM
22
23
0
23 Jun 2023
Towards Explainable Evaluation Metrics for Machine Translation
Towards Explainable Evaluation Metrics for Machine Translation
Christoph Leiter
Piyawat Lertvittayakumjorn
M. Fomicheva
Wei-Ye Zhao
Yang Gao
Steffen Eger
ELM
28
12
0
22 Jun 2023
Towards Regulatable AI Systems: Technical Gaps and Policy Opportunities
Towards Regulatable AI Systems: Technical Gaps and Policy Opportunities
Xudong Shen
H. Brown
Jiashu Tao
Martin Strobel
Yao Tong
Akshay Narayan
Harold Soh
Finale Doshi-Velez
27
3
0
22 Jun 2023
Which Spurious Correlations Impact Reasoning in NLI Models? A Visual
  Interactive Diagnosis through Data-Constrained Counterfactuals
Which Spurious Correlations Impact Reasoning in NLI Models? A Visual Interactive Diagnosis through Data-Constrained Counterfactuals
Robin Shing Moon Chan
Afra Amini
Mennatallah El-Assady
LRM
AAML
29
2
0
21 Jun 2023
Causal Effect Regularization: Automated Detection and Removal of
  Spurious Attributes
Causal Effect Regularization: Automated Detection and Removal of Spurious Attributes
Abhinav Kumar
Amit Deshpande
Ajay Sharma
CML
11
1
0
19 Jun 2023
Cross-Modal Attribute Insertions for Assessing the Robustness of
  Vision-and-Language Learning
Cross-Modal Attribute Insertions for Assessing the Robustness of Vision-and-Language Learning
Shivaen Ramshetty
Gaurav Verma
Srijan Kumar
30
2
0
19 Jun 2023
Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis,
  and LLMs Evaluations
Revisiting Out-of-distribution Robustness in NLP: Benchmark, Analysis, and LLMs Evaluations
Lifan Yuan
Yangyi Chen
Ganqu Cui
Hongcheng Gao
Fangyuan Zou
Xingyi Cheng
Heng Ji
Zhiyuan Liu
Maosong Sun
32
72
0
07 Jun 2023
Reason to explain: Interactive contrastive explanations (REASONX)
Reason to explain: Interactive contrastive explanations (REASONX)
Laura State
Salvatore Ruggieri
Franco Turini
LRM
22
1
0
29 May 2023
Faithfulness Tests for Natural Language Explanations
Faithfulness Tests for Natural Language Explanations
Pepa Atanasova
Oana-Maria Camburu
Christina Lioma
Thomas Lukasiewicz
J. Simonsen
Isabelle Augenstein
FAtt
21
59
0
29 May 2023
CREST: A Joint Framework for Rationalization and Counterfactual Text
  Generation
CREST: A Joint Framework for Rationalization and Counterfactual Text Generation
Marcos Vinícius Treviso
Alexis Ross
Nuno M. Guerreiro
André F.T. Martins
29
16
0
26 May 2023
Counterfactuals of Counterfactuals: a back-translation-inspired approach
  to analyse counterfactual editors
Counterfactuals of Counterfactuals: a back-translation-inspired approach to analyse counterfactual editors
Giorgos Filandrianos
Edmund Dervakos
Orfeas Menis-Mastromichalakis
Chrysoula Zerva
Giorgos Stamou
AAML
24
4
0
26 May 2023
Controlling Learned Effects to Reduce Spurious Correlations in Text
  Classifiers
Controlling Learned Effects to Reduce Spurious Correlations in Text Classifiers
Parikshit Bansal
Amit Sharma
CML
24
5
0
26 May 2023
On Degrees of Freedom in Defining and Testing Natural Language
  Understanding
On Degrees of Freedom in Defining and Testing Natural Language Understanding
Saku Sugawara
S. Tsugita
ELM
21
1
0
24 May 2023
Clever Hans or Neural Theory of Mind? Stress Testing Social Reasoning in
  Large Language Models
Clever Hans or Neural Theory of Mind? Stress Testing Social Reasoning in Large Language Models
Natalie Shapira
Mosh Levy
S. Alavi
Xuhui Zhou
Yejin Choi
Yoav Goldberg
Maarten Sap
Vered Shwartz
LLMAG
ELM
20
113
0
24 May 2023
Exploring Contrast Consistency of Open-Domain Question Answering Systems
  on Minimally Edited Questions
Exploring Contrast Consistency of Open-Domain Question Answering Systems on Minimally Edited Questions
Zhihan Zhang
W. Yu
Zheng Ning
Mingxuan Ju
Meng-Long Jiang
16
4
0
23 May 2023
Out-of-Distribution Generalization in Text Classification: Past,
  Present, and Future
Out-of-Distribution Generalization in Text Classification: Past, Present, and Future
Linyi Yang
Y. Song
Xuan Ren
Chenyang Lyu
Yidong Wang
Lingqiao Liu
Jindong Wang
Jennifer Foster
Yue Zhang
OOD
20
2
0
23 May 2023
Improving Classifier Robustness through Active Generation of Pairwise
  Counterfactuals
Improving Classifier Robustness through Active Generation of Pairwise Counterfactuals
Ananth Balashankar
Xuezhi Wang
Yao Qin
Ben Packer
Nithum Thain
Jilin Chen
Ed H. Chi
Alex Beutel
17
0
0
22 May 2023
Previous
1234
Next