ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2204.03025
  4. Cited By
Using Interactive Feedback to Improve the Accuracy and Explainability of
  Question Answering Systems Post-Deployment

Using Interactive Feedback to Improve the Accuracy and Explainability of Question Answering Systems Post-Deployment

6 April 2022
Zichao Li
Prakhar Sharma
Xing Han Lù
Jackie C.K. Cheung
Siva Reddy
    HAI
ArXivPDFHTML

Papers citing "Using Interactive Feedback to Improve the Accuracy and Explainability of Question Answering Systems Post-Deployment"

20 / 20 papers shown
Title
Aligning Black-box Language Models with Human Judgments
Aligning Black-box Language Models with Human Judgments
Gerrit J. J. van den Burg
Gen Suzuki
Wei Liu
Murat Sensoy
ALM
82
0
0
07 Feb 2025
Black-box Uncertainty Quantification Method for LLM-as-a-Judge
Black-box Uncertainty Quantification Method for LLM-as-a-Judge
Nico Wagner
Michael Desmond
Rahul Nair
Zahra Ashktorab
Elizabeth M. Daly
Qian Pan
Martin Santillan Cooper
James M. Johnson
Werner Geyer
ELM
UQCV
44
4
0
15 Oct 2024
Honesty to Subterfuge: In-Context Reinforcement Learning Can Make Honest
  Models Reward Hack
Honesty to Subterfuge: In-Context Reinforcement Learning Can Make Honest Models Reward Hack
Leo McKee-Reid
Christoph Sträter
Maria Angelica Martinez
Joe Needham
Mikita Balesni
OffRL
32
2
0
09 Oct 2024
Aligning LLM Agents by Learning Latent Preference from User Edits
Aligning LLM Agents by Learning Latent Preference from User Edits
Ge Gao
Alexey Taymanov
Eduardo Salinas
Paul Mineiro
Dipendra Kumar Misra
LLMAG
37
27
0
23 Apr 2024
RLHF Deciphered: A Critical Analysis of Reinforcement Learning from
  Human Feedback for LLMs
RLHF Deciphered: A Critical Analysis of Reinforcement Learning from Human Feedback for LLMs
Shreyas Chaudhari
Pranjal Aggarwal
Vishvak Murahari
Tanmay Rajpurohit
A. Kalyan
Karthik Narasimhan
A. Deshpande
Bruno Castro da Silva
23
34
0
12 Apr 2024
Are self-explanations from Large Language Models faithful?
Are self-explanations from Large Language Models faithful?
Andreas Madsen
Sarath Chandar
Siva Reddy
LRM
30
24
0
15 Jan 2024
Reasons to Reject? Aligning Language Models with Judgments
Reasons to Reject? Aligning Language Models with Judgments
Weiwen Xu
Deng Cai
Zhisong Zhang
Wai Lam
Shuming Shi
ALM
19
14
0
22 Dec 2023
What if you said that differently?: How Explanation Formats Affect Human
  Feedback Efficacy and User Perception
What if you said that differently?: How Explanation Formats Affect Human Feedback Efficacy and User Perception
Chaitanya Malaviya
Subin Lee
Dan Roth
Mark Yatskar
31
1
0
16 Nov 2023
Continually Improving Extractive QA via Human Feedback
Continually Improving Extractive QA via Human Feedback
Ge Gao
Hung-Ting Chen
Yoav Artzi
Eunsol Choi
24
12
0
21 May 2023
Learning to Simulate Natural Language Feedback for Interactive Semantic
  Parsing
Learning to Simulate Natural Language Feedback for Interactive Semantic Parsing
Hao Yan
Saurabh Srivastava
Yintao Tai
Sida I. Wang
Wen-tau Yih
Ziyu Yao
22
17
0
14 May 2023
Bridging the Gap: A Survey on Integrating (Human) Feedback for Natural
  Language Generation
Bridging the Gap: A Survey on Integrating (Human) Feedback for Natural Language Generation
Patrick Fernandes
Aman Madaan
Emmy Liu
António Farinhas
Pedro Henrique Martins
...
José G. C. de Souza
Shuyan Zhou
Tongshuang Wu
Graham Neubig
André F. T. Martins
ALM
117
56
0
01 May 2023
Training Language Models with Language Feedback at Scale
Training Language Models with Language Feedback at Scale
Jérémy Scheurer
Jon Ander Campos
Tomasz Korbak
Jun Shern Chan
Angelica Chen
Kyunghyun Cho
Ethan Perez
ALM
39
101
0
28 Mar 2023
MWE as WSD: Solving Multiword Expression Identification with Word Sense
  Disambiguation
MWE as WSD: Solving Multiword Expression Identification with Word Sense Disambiguation
Joshua Tanner
J. Hoffman
16
1
0
12 Mar 2023
Robust Question Answering against Distribution Shifts with Test-Time
  Adaptation: An Empirical Study
Robust Question Answering against Distribution Shifts with Test-Time Adaptation: An Empirical Study
Hai Ye
Yuyang Ding
Juntao Li
Hwee Tou Ng
OOD
TTA
21
9
0
09 Feb 2023
InterFair: Debiasing with Natural Language Feedback for Fair
  Interpretable Predictions
InterFair: Debiasing with Natural Language Feedback for Fair Interpretable Predictions
Bodhisattwa Prasad Majumder
Zexue He
Julian McAuley
6
5
0
14 Oct 2022
Training Language Models with Language Feedback
Training Language Models with Language Feedback
Jérémy Scheurer
Jon Ander Campos
Jun Shern Chan
Angelica Chen
Kyunghyun Cho
Ethan Perez
ALM
36
47
0
29 Apr 2022
Local Interpretations for Explainable Natural Language Processing: A
  Survey
Local Interpretations for Explainable Natural Language Processing: A Survey
Siwen Luo
Hamish Ivison
S. Han
Josiah Poon
MILM
33
48
0
20 Mar 2021
ExpBERT: Representation Engineering with Natural Language Explanations
ExpBERT: Representation Engineering with Natural Language Explanations
Shikhar Murty
Pang Wei Koh
Percy Liang
38
43
0
05 May 2020
e-SNLI: Natural Language Inference with Natural Language Explanations
e-SNLI: Natural Language Inference with Natural Language Explanations
Oana-Maria Camburu
Tim Rocktaschel
Thomas Lukasiewicz
Phil Blunsom
LRM
255
620
0
04 Dec 2018
Dialogue Learning With Human-In-The-Loop
Dialogue Learning With Human-In-The-Loop
Jiwei Li
Alexander H. Miller
S. Chopra
MarcÁurelio Ranzato
Jason Weston
OffRL
218
134
0
29 Nov 2016
1