ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2004.11440
  4. Cited By
Human Factors in Model Interpretability: Industry Practices, Challenges,
  and Needs

Human Factors in Model Interpretability: Industry Practices, Challenges, and Needs

23 April 2020
Sungsoo Ray Hong
Jessica Hullman
E. Bertini
    HAI
ArXivPDFHTML

Papers citing "Human Factors in Model Interpretability: Industry Practices, Challenges, and Needs"

50 / 89 papers shown
Title
Evolutionary Reinforcement Learning for Interpretable Decision-Making in Supply Chain Management
Evolutionary Reinforcement Learning for Interpretable Decision-Making in Supply Chain Management
Stefano Genetti
Alberto Longobardi
Giovanni Iacca
61
0
0
16 Apr 2025
SPHERE: An Evaluation Card for Human-AI Systems
SPHERE: An Evaluation Card for Human-AI Systems
Qianou Ma
Dora Zhao
Xinran Zhao
Chenglei Si
Chenyang Yang
Ryan Louie
Ehud Reiter
Diyi Yang
Tongshuang Wu
ALM
68
1
0
24 Mar 2025
EXAGREE: Towards Explanation Agreement in Explainable Machine Learning
EXAGREE: Towards Explanation Agreement in Explainable Machine Learning
Sichao Li
Quanling Deng
Amanda S. Barnard
47
0
0
04 Nov 2024
Towards Human-centered Design of Explainable Artificial Intelligence
  (XAI): A Survey of Empirical Studies
Towards Human-centered Design of Explainable Artificial Intelligence (XAI): A Survey of Empirical Studies
Shuai Ma
34
0
0
28 Oct 2024
Behavior Matters: An Alternative Perspective on Promoting Responsible
  Data Science
Behavior Matters: An Alternative Perspective on Promoting Responsible Data Science
Ziwei Dong
Ameya Patil
Yuichi Shoda
Leilani Battle
Emily Wall
AI4CE
35
1
0
07 Oct 2024
The FIX Benchmark: Extracting Features Interpretable to eXperts
The FIX Benchmark: Extracting Features Interpretable to eXperts
Helen Jin
Shreya Havaldar
Chaehyeon Kim
Anton Xue
Weiqiu You
...
Bhuvnesh Jain
Amin Madani
M. Sako
Lyle Ungar
Eric Wong
31
1
0
20 Sep 2024
The Atlas of AI Incidents in Mobile Computing: Visualizing the Risks and
  Benefits of AI Gone Mobile
The Atlas of AI Incidents in Mobile Computing: Visualizing the Risks and Benefits of AI Gone Mobile
Edyta Bogucka
Marios Constantinides
Julia De Miguel Velazquez
S. Šćepanović
Daniele Quercia
Andrés Gvirtz
47
1
0
22 Jul 2024
Operationalizing the Blueprint for an AI Bill of Rights: Recommendations
  for Practitioners, Researchers, and Policy Makers
Operationalizing the Blueprint for an AI Bill of Rights: Recommendations for Practitioners, Researchers, and Policy Makers
Alex Oesterling
Usha Bhalla
Suresh Venkatasubramanian
Himabindu Lakkaraju
56
1
0
11 Jul 2024
"It answers questions that I didn't know I had": Ph.D. Students'
  Evaluation of an Information Sharing Knowledge Graph
"It answers questions that I didn't know I had": Ph.D. Students' Evaluation of an Information Sharing Knowledge Graph
S. Gardasevic
Manika Lamba
33
1
0
11 Jun 2024
Explainable Interface for Human-Autonomy Teaming: A Survey
Explainable Interface for Human-Autonomy Teaming: A Survey
Xiangqi Kong
Yang Xing
Antonios Tsourdos
Ziyue Wang
Weisi Guo
Adolfo Perrusquía
Andreas Wikander
48
3
0
04 May 2024
Exploring Algorithmic Explainability: Generating Explainable AI Insights
  for Personalized Clinical Decision Support Focused on Cannabis Intoxication
  in Young Adults
Exploring Algorithmic Explainability: Generating Explainable AI Insights for Personalized Clinical Decision Support Focused on Cannabis Intoxication in Young Adults
Tongze Zhang
Tammy Chung
Anind Dey
Sang Won Bae
36
3
0
22 Apr 2024
Demonstration of MaskSearch: Efficiently Querying Image Masks for
  Machine Learning Workflows
Demonstration of MaskSearch: Efficiently Querying Image Masks for Machine Learning Workflows
Lindsey Linxi Wei
Chung Yik Edward Yeung
Hongjian Yu
Jingchuan Zhou
Dong He
Magdalena Balazinska
OOD
26
0
0
09 Apr 2024
Explainability in JupyterLab and Beyond: Interactive XAI Systems for
  Integrated and Collaborative Workflows
Explainability in JupyterLab and Beyond: Interactive XAI Systems for Integrated and Collaborative Workflows
G. Guo
Dustin L. Arendt
Alex Endert
53
1
0
02 Apr 2024
A Preliminary Roadmap for LLMs as Assistants in Exploring, Analyzing,
  and Visualizing Knowledge Graphs
A Preliminary Roadmap for LLMs as Assistants in Exploring, Analyzing, and Visualizing Knowledge Graphs
Harry Li
G. Appleby
Ashley Suh
47
7
0
01 Apr 2024
Closing the Knowledge Gap in Designing Data Annotation Interfaces for
  AI-powered Disaster Management Analytic Systems
Closing the Knowledge Gap in Designing Data Annotation Interfaces for AI-powered Disaster Management Analytic Systems
Zinat Ara
Hossein Salemi
Sungsoo Ray Hong
Yasas Senarath
Steve Peterson
A. Hughes
Hemant Purohit
45
2
0
04 Mar 2024
Content-Centric Prototyping of Generative AI Applications: Emerging
  Approaches and Challenges in Collaborative Software Teams
Content-Centric Prototyping of Generative AI Applications: Emerging Approaches and Challenges in Collaborative Software Teams
Hari Subramonyam
Divy Thakkar
Jurgen Dieber
Anoop Sinha
55
3
0
27 Feb 2024
LCEN: A Novel Feature Selection Algorithm for Nonlinear, Interpretable
  Machine Learning Models
LCEN: A Novel Feature Selection Algorithm for Nonlinear, Interpretable Machine Learning Models
Pedro Seber
R. D. Braatz
44
3
0
27 Feb 2024
Farsight: Fostering Responsible AI Awareness During AI Application
  Prototyping
Farsight: Fostering Responsible AI Awareness During AI Application Prototyping
Zijie J. Wang
Chinmay Kulkarni
Lauren Wilcox
Michael Terry
Michael A. Madaio
45
44
0
23 Feb 2024
NormEnsembleXAI: Unveiling the Strengths and Weaknesses of XAI Ensemble
  Techniques
NormEnsembleXAI: Unveiling the Strengths and Weaknesses of XAI Ensemble Techniques
Weronika Hryniewska-Guzik
Bartosz Sawicki
P. Biecek
38
0
0
30 Jan 2024
3DPFIX: Improving Remote Novices' 3D Printing Troubleshooting through
  Human-AI Collaboration
3DPFIX: Improving Remote Novices' 3D Printing Troubleshooting through Human-AI Collaboration
Nahyun Kwon
Tong Sun
Yuyang Gao
Liang Zhao
Xu Wang
Jeeeun Kim
S. Hong
21
2
0
29 Jan 2024
Towards a Non-Ideal Methodological Framework for Responsible ML
Towards a Non-Ideal Methodological Framework for Responsible ML
Ramaravind Kommiya Mothilal
Shion Guha
Syed Ishtiaque Ahmed
59
7
0
20 Jan 2024
Pyreal: A Framework for Interpretable ML Explanations
Pyreal: A Framework for Interpretable ML Explanations
Alexandra Zytek
Wei-En Wang
Dongyu Liu
Laure Berti-Equille
K. Veeramachaneni
LRM
50
0
0
20 Dec 2023
Lessons from Usable ML Deployments and Application to Wind Turbine
  Monitoring
Lessons from Usable ML Deployments and Application to Wind Turbine Monitoring
Alexandra Zytek
Wei-En Wang
S. Koukoura
K. Veeramachaneni
50
0
0
05 Dec 2023
Model Compression in Practice: Lessons Learned from Practitioners
  Creating On-device Machine Learning Experiences
Model Compression in Practice: Lessons Learned from Practitioners Creating On-device Machine Learning Experiences
Fred Hohman
Mary Beth Kery
Donghao Ren
Dominik Moritz
37
16
0
06 Oct 2023
Personalized Guidelines for Design, Implementation and Evaluation of
  Anti-phishing Interventions
Personalized Guidelines for Design, Implementation and Evaluation of Anti-phishing Interventions
Orvila Sarker
Sherif Haggag
Asangi Jayatilaka
Chelsea Liu
25
3
0
01 Oct 2023
A Survey on Image-text Multimodal Models
A Survey on Image-text Multimodal Models
Ruifeng Guo
Jingxuan Wei
Linzhuang Sun
Khai Le-Duc
Guiyong Chang
Dawei Liu
Sibo Zhang
Zhengbing Yao
Mingjun Xu
Liping Bu
VLM
36
5
0
23 Sep 2023
TRIVEA: Transparent Ranking Interpretation using Visual Explanation of
  Black-Box Algorithmic Rankers
TRIVEA: Transparent Ranking Interpretation using Visual Explanation of Black-Box Algorithmic Rankers
Jun Yuan
Kaustav Bhattacharjee
A. Islam
Aritra Dasgupta
28
2
0
28 Aug 2023
Comparing AutoML and Deep Learning Methods for Condition Monitoring
  using Realistic Validation Scenarios
Comparing AutoML and Deep Learning Methods for Condition Monitoring using Realistic Validation Scenarios
P. Goodarzi
A. Schütze
T. Schneider
31
0
0
28 Aug 2023
Are We Closing the Loop Yet? Gaps in the Generalizability of VIS4ML
  Research
Are We Closing the Loop Yet? Gaps in the Generalizability of VIS4ML Research
Hariharan Subramonyam
Jessica Hullman
VLM
HAI
32
8
0
10 Aug 2023
Identifying Explanation Needs of End-users: Applying and Extending the
  XAI Question Bank
Identifying Explanation Needs of End-users: Applying and Extending the XAI Question Bank
Lars Sipos
Ulrike Schäfer
Katrin Glinka
Claudia Muller-Birn
33
7
0
18 Jul 2023
Designing a Direct Feedback Loop between Humans and Convolutional Neural
  Networks through Local Explanations
Designing a Direct Feedback Loop between Humans and Convolutional Neural Networks through Local Explanations
Tong Sun
Yuyang Gao
Shubham Khaladkar
Sijia Liu
Liang Zhao
Younghoon Kim
S. Hong
AAML
FAtt
HAI
34
6
0
08 Jul 2023
AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap
AI Transparency in the Age of LLMs: A Human-Centered Research Roadmap
Q. V. Liao
J. Vaughan
58
159
0
02 Jun 2023
Rethinking Model Evaluation as Narrowing the Socio-Technical Gap
Rethinking Model Evaluation as Narrowing the Socio-Technical Gap
Q. V. Liao
Ziang Xiao
ALM
ELM
66
30
0
01 Jun 2023
MaskSearch: Querying Image Masks at Scale
MaskSearch: Querying Image Masks at Scale
Dong He
Jieyu Zhang
Maureen Daum
Alexander Ratner
Magdalena Balazinska
VLM
42
2
0
03 May 2023
Tracing and Visualizing Human-ML/AI Collaborative Processes through
  Artifacts of Data Work
Tracing and Visualizing Human-ML/AI Collaborative Processes through Artifacts of Data Work
Jennifer Rogers
Anamaria Crisan
40
7
0
05 Apr 2023
Knowledge Graphs in Practice: Characterizing their Users, Challenges,
  and Visualization Opportunities
Knowledge Graphs in Practice: Characterizing their Users, Challenges, and Visualization Opportunities
Harry Li
G. Appleby
C. Brumar
Remco Chang
Ashley Suh
31
23
0
03 Apr 2023
Designerly Understanding: Information Needs for Model Transparency to
  Support Design Ideation for AI-Powered User Experience
Designerly Understanding: Information Needs for Model Transparency to Support Design Ideation for AI-Powered User Experience
Q. V. Liao
Hariharan Subramonyam
Jennifer Wang
Jennifer Wortman Vaughan
HAI
38
59
0
21 Feb 2023
Invisible Users: Uncovering End-Users' Requirements for Explainable AI
  via Explanation Forms and Goals
Invisible Users: Uncovering End-Users' Requirements for Explainable AI via Explanation Forms and Goals
Weina Jin
Jianyu Fan
D. Gromala
Philippe Pasquier
Ghassan Hamarneh
36
7
0
10 Feb 2023
Five policy uses of algorithmic transparency and explainability
Five policy uses of algorithmic transparency and explainability
Matthew R. O’Shaughnessy
59
0
0
06 Feb 2023
Ignore, Trust, or Negotiate: Understanding Clinician Acceptance of
  AI-Based Treatment Recommendations in Health Care
Ignore, Trust, or Negotiate: Understanding Clinician Acceptance of AI-Based Treatment Recommendations in Health Care
Venkatesh Sivaraman
L. Bukowski
J. Levin
J. Kahn
Adam Perer
40
83
0
31 Jan 2023
Towards Multifaceted Human-Centered AI
Towards Multifaceted Human-Centered AI
Sajjadur Rahman
H. Kim
Dan Zhang
Estevam R. Hruschka
Eser Kandogan
18
1
0
09 Jan 2023
Going Beyond XAI: A Systematic Survey for Explanation-Guided Learning
Going Beyond XAI: A Systematic Survey for Explanation-Guided Learning
Yuyang Gao
Siyi Gu
Junji Jiang
S. Hong
Dazhou Yu
Liang Zhao
31
39
0
07 Dec 2022
Explainable Artificial Intelligence (XAI) from a user perspective- A
  synthesis of prior literature and problematizing avenues for future research
Explainable Artificial Intelligence (XAI) from a user perspective- A synthesis of prior literature and problematizing avenues for future research
Akm Bahalul Haque
A. Najmul Islam
Patrick Mikalef
XAI
32
95
0
24 Nov 2022
Seamful XAI: Operationalizing Seamful Design in Explainable AI
Seamful XAI: Operationalizing Seamful Design in Explainable AI
Upol Ehsan
Q. V. Liao
Samir Passi
Mark O. Riedl
Hal Daumé
30
20
0
12 Nov 2022
Black Box Model Explanations and the Human Interpretability Expectations
  -- An Analysis in the Context of Homicide Prediction
Black Box Model Explanations and the Human Interpretability Expectations -- An Analysis in the Context of Homicide Prediction
José Ribeiro
Nikolas Carneiro
Ronnie Cley de Oliveira Alves
19
0
0
19 Oct 2022
A.I. Robustness: a Human-Centered Perspective on Technological
  Challenges and Opportunities
A.I. Robustness: a Human-Centered Perspective on Technological Challenges and Opportunities
Andrea Tocchetti
Lorenzo Corti
Agathe Balayn
Mireia Yurrita
Philip Lippmann
Marco Brambilla
Jie Yang
32
10
0
17 Oct 2022
Self-explaining deep models with logic rule reasoning
Self-explaining deep models with logic rule reasoning
Seungeon Lee
Xiting Wang
Sungwon Han
Xiaoyuan Yi
Xing Xie
M. Cha
NAI
ReLM
LRM
40
16
0
13 Oct 2022
"Help Me Help the AI": Understanding How Explainability Can Support
  Human-AI Interaction
"Help Me Help the AI": Understanding How Explainability Can Support Human-AI Interaction
Sunnie S. Y. Kim
E. A. Watkins
Olga Russakovsky
Ruth C. Fong
Andrés Monroy-Hernández
43
108
0
02 Oct 2022
How Do We Measure Trust in Visual Data Communication?
How Do We Measure Trust in Visual Data Communication?
Hamza Elhamdadi
Aimen Gaba
Yea-Seul Kim
Cindy Xiong
20
18
0
28 Sep 2022
"Is It My Turn?" Assessing Teamwork and Taskwork in Collaborative
  Immersive Analytics
"Is It My Turn?" Assessing Teamwork and Taskwork in Collaborative Immersive Analytics
Michaela Benk
Raphael P. Weibel
Stefan Feuerriegel
Andrea Ferrario
31
3
0
09 Aug 2022
12
Next