ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Communities
  3. ...

Neighbor communities

0 / 0 papers shown
Title
Top Contributors
Name# Papers# Citations
Social Events
DateLocationEvent
  1. Home
  2. Communities
  3. HILM

Hallucination in Language Models

HILM
More data

Dedicated to studies primarily investigating the causes, implications, and solutions for the phenomenon where language models generate plausible but incorrect or nonsensical outputs.

Neighbor communities

51015

Featured Papers

0 / 0 papers shown
Title

All papers

50 / 1,186 papers shown
Title
When Bias Pretends to Be Truth: How Spurious Correlations Undermine Hallucination Detection in LLMs
When Bias Pretends to Be Truth: How Spurious Correlations Undermine Hallucination Detection in LLMs
Shaowen Wang
Yiqi Dong
Ruinian Chang
Tansheng Zhu
Yuebo Sun
Kaifeng Lyu
Jian Li
HILM
0
0
0
10 Nov 2025
When Evidence Contradicts: Toward Safer Retrieval-Augmented Generation in Healthcare
When Evidence Contradicts: Toward Safer Retrieval-Augmented Generation in Healthcare
Saeedeh Javadi
Sara Mirabi
Manan Gangar
Bahadorreza Ofoghi
RALMHILM
4
0
0
10 Nov 2025
Stress Testing Factual Consistency Metrics for Long-Document Summarization
Stress Testing Factual Consistency Metrics for Long-Document Summarization
Zain Muhammad Mujahid
Dustin Wright
Isabelle Augenstein
HILM
0
0
0
10 Nov 2025
NOAH: Benchmarking Narrative Prior driven Hallucination and Omission in Video Large Language Models
NOAH: Benchmarking Narrative Prior driven Hallucination and Omission in Video Large Language Models
Kyuho Lee
Euntae Kim
Jinwoo Choi
Buru Chang
HILM
0
0
0
09 Nov 2025
Injecting Falsehoods: Adversarial Man-in-the-Middle Attacks Undermining Factual Recall in LLMs
Injecting Falsehoods: Adversarial Man-in-the-Middle Attacks Undermining Factual Recall in LLMs
Alina Fastowski
Bardh Prenkaj
Yuxiao Li
Gjergji Kasneci
AAMLKELMHILM
80
0
0
08 Nov 2025
Stemming Hallucination in Language Models Using a Licensing Oracle
Stemming Hallucination in Language Models Using a Licensing Oracle
Simeon Emanuilov
Richard Ackermann
HILM
0
0
0
08 Nov 2025
REFLEX: Reference-Free Evaluation of Log Summarization via Large Language Model Judgment
REFLEX: Reference-Free Evaluation of Log Summarization via Large Language Model Judgment
Priyanka Mudgal
HILM
68
0
0
06 Nov 2025
HaluMem: Evaluating Hallucinations in Memory Systems of Agents
HaluMem: Evaluating Hallucinations in Memory Systems of Agents
Ding Chen
Simin Niu
Kehang Li
Peng Liu
Xiangping Zheng
Bo Tang
X. Li
Feiyu Xiong
Zhiyu Li
LLMAGHILMVLM
122
0
0
05 Nov 2025
PrefixNLI: Detecting Factual Inconsistencies as Soon as They Arise
PrefixNLI: Detecting Factual Inconsistencies as Soon as They Arise
Sapir Harary
Eran Hirsch
Aviv Slobodkin
David Wan
Mohit Bansal
Ido Dagan
HILM
149
0
0
03 Nov 2025
VISTA Score: Verification In Sequential Turn-based Assessment
VISTA Score: Verification In Sequential Turn-based Assessment
A. Lewis
Andrew Perrault
Eric Fosler-Lussier
Michael White
HILM
117
0
0
30 Oct 2025
SciTrust 2.0: A Comprehensive Framework for Evaluating Trustworthiness of Large Language Models in Scientific Applications
SciTrust 2.0: A Comprehensive Framework for Evaluating Trustworthiness of Large Language Models in Scientific Applications
Emily Herron
Junqi Yin
Feiyi Wang
HILMELM
151
0
0
29 Oct 2025
Layer of Truth: Probing Belief Shifts under Continual Pre-Training Poisoning
Layer of Truth: Probing Belief Shifts under Continual Pre-Training Poisoning
S. Churina
Niranjan Chebrolu
Kokil Jaidka
KELMHILMCLL
160
0
0
29 Oct 2025
Hallucinations in Bibliographic Recommendation: Citation Frequency as a Proxy for Training Data Redundancy
Hallucinations in Bibliographic Recommendation: Citation Frequency as a Proxy for Training Data Redundancy
Junichiro Niimi
HILMRALM
68
0
0
29 Oct 2025
Challenging Multilingual LLMs: A New Taxonomy and Benchmark for Unraveling Hallucination in Translation
Challenging Multilingual LLMs: A New Taxonomy and Benchmark for Unraveling Hallucination in Translation
Xinwei Wu
Heng Liu
Jiang Zhou
Xiaohu Zhao
Linlong Xu
Longyue Wang
Weihua Luo
Kaifu Zhang
HILM
124
0
0
28 Oct 2025
MAD-Fact: A Multi-Agent Debate Framework for Long-Form Factuality Evaluation in LLMs
MAD-Fact: A Multi-Agent Debate Framework for Long-Form Factuality Evaluation in LLMs
Yucheng Ning
Xixun Lin
Fang Fang
Yanan Cao
HILM
117
0
0
27 Oct 2025
MR-Align: Meta-Reasoning Informed Factuality Alignment for Large Reasoning Models
MR-Align: Meta-Reasoning Informed Factuality Alignment for Large Reasoning Models
Xinming Wang
Jian Xu
Bin Yu
Sheng Lian
Hongzhu Yi
...
Boran Wang
Hongming Yang
Han Hu
Xu-Yao Zhang
Cheng-Lin Liu
HILMLRM
68
0
0
27 Oct 2025
Multi-Modal Fact-Verification Framework for Reducing Hallucinations in Large Language Models
Multi-Modal Fact-Verification Framework for Reducing Hallucinations in Large Language Models
Piyushkumar Patel
HILM
68
0
0
26 Oct 2025
Confabulations from ACL Publications (CAP): A Dataset for Scientific Hallucination Detection
Confabulations from ACL Publications (CAP): A Dataset for Scientific Hallucination Detection
Federica Gamba
Aman Sinha
Timothee Mickus
Raul Vazquez
Patanjali Bhamidipati
...
Aryan Chandramania
Rohit Agarwal
Chuyuan Li
Ioana Buhnila
Radhika Mamidi
HILM
56
0
0
25 Oct 2025
The Gray Zone of Faithfulness: Taming Ambiguity in Unfaithfulness Detection
The Gray Zone of Faithfulness: Taming Ambiguity in Unfaithfulness Detection
Qiang Ding
Lvzhou Luo
Yixuan Cao
Ping Luo
HILM
117
0
0
24 Oct 2025
Embedding Trust: Semantic Isotropy Predicts Nonfactuality in Long-Form Text Generation
Embedding Trust: Semantic Isotropy Predicts Nonfactuality in Long-Form Text Generation
Dhrupad Bhardwaj
Julia Kempe
Tim G. J. Rudner
HILM
115
0
0
24 Oct 2025
A Diagnostic Benchmark for Sweden-Related Factual Knowledge
A Diagnostic Benchmark for Sweden-Related Factual Knowledge
Jenny Kunz
HILM
52
0
0
24 Oct 2025
A Benchmark for Open-Domain Numerical Fact-Checking Enhanced by Claim Decomposition
A Benchmark for Open-Domain Numerical Fact-Checking Enhanced by Claim Decomposition
Venktesh V
Deepali Prabhu
Avishek Anand
HILM
63
0
0
24 Oct 2025
Teaming LLMs to Detect and Mitigate Hallucinations
Teaming LLMs to Detect and Mitigate Hallucinations
Demian Till
John Smeaton
Peter Haubrick
Gouse Saheb
Florian Graef
David Berman
HILM
123
0
0
22 Oct 2025
JointCQ: Improving Factual Hallucination Detection with Joint Claim and Query Generation
JointCQ: Improving Factual Hallucination Detection with Joint Claim and Query Generation
F. Xu
Huixuan Zhang
Zhenliang Zhang
Jiahao Wang
Xiaojun Wan
HILM
80
0
0
22 Oct 2025
HAD: HAllucination Detection Language Models Based on a Comprehensive Hallucination Taxonomy
HAD: HAllucination Detection Language Models Based on a Comprehensive Hallucination Taxonomy
Fan Xu
Xinyu Hu
Zhenghan Yu
Li Lin
Xu Zhang
Yang Zhang
Wei Zhou
Jinjie Gu
Xiaojun Wan
HILM
60
0
0
22 Oct 2025
KoSimpleQA: A Korean Factuality Benchmark with an Analysis of Reasoning LLMs
KoSimpleQA: A Korean Factuality Benchmark with an Analysis of Reasoning LLMs
Donghyeon Ko
Yeguk Jin
Kyubyung Chae
Byungwook Lee
Chansong Jo
Sookyo In
Jaehong Lee
Taesup Kim
Donghyun Kwak
HILM
60
0
0
21 Oct 2025
JT-Safe: Intrinsically Enhancing the Safety and Trustworthiness of LLMs
JT-Safe: Intrinsically Enhancing the Safety and Trustworthiness of LLMs
Junlan Feng
Fanyu Meng
Chong Long
Pengyu Cong
Duqing Wang
...
Z. Ren
Fan Yang
Na Wu
Di Jin
Chao Deng
HILM
78
0
0
20 Oct 2025
Train for Truth, Keep the Skills: Binary Retrieval-Augmented Reward Mitigates Hallucinations
Train for Truth, Keep the Skills: Binary Retrieval-Augmented Reward Mitigates Hallucinations
Tong Chen
Akari Asai
Luke Zettlemoyer
Hannaneh Hajishirzi
Faeze Brahman
OffRLHILMLRM
64
0
0
20 Oct 2025
Annotation-Efficient Universal Honesty Alignment
Annotation-Efficient Universal Honesty Alignment
Shiyu Ni
Keping Bi
Jiafeng Guo
Minghao Tang
Jingtong Wu
Zengxin Han
Xueqi Cheng
HILM
48
0
0
20 Oct 2025
Hallucination Benchmark for Speech Foundation Models
Hallucination Benchmark for Speech Foundation Models
Alkis Koudounas
Moreno La Quatra
Manuel Giollo
Sabato Marco Siniscalchi
Elena Baralis
HILM
86
0
0
18 Oct 2025
MedTrust-RAG: Evidence Verification and Trust Alignment for Biomedical Question Answering
MedTrust-RAG: Evidence Verification and Trust Alignment for Biomedical Question Answering
Yingpeng Ning
Yuanyuan Sun
Ling Luo
Yanhua Wang
Yuchen Pan
Hongfei Lin
HILM
140
0
0
16 Oct 2025
Counting Hallucinations in Diffusion Models
Counting Hallucinations in Diffusion Models
Shuai Fu
Jian Zhou
Qi Chen
Huang Jing
Huy Anh Nguyen
Xiaohan Liu
Zhixiong Zeng
Lin Ma
Quanshi Zhang
Qi Wu
DiffMHILM
99
0
0
15 Oct 2025
Teaching Language Models to Faithfully Express their Uncertainty
Teaching Language Models to Faithfully Express their Uncertainty
Bryan Eikema
Evgenia Ilia
José G. C. de Souza
Chrysoula Zerva
Wilker Aziz
HILM
48
0
0
14 Oct 2025
Credal Transformer: A Principled Approach for Quantifying and Mitigating Hallucinations in Large Language Models
Credal Transformer: A Principled Approach for Quantifying and Mitigating Hallucinations in Large Language Models
Shihao Ji
Zihui Song
Jiajie Huang
HILM
48
0
0
14 Oct 2025
Uncertainty Quantification for Hallucination Detection in Large Language Models: Foundations, Methodology, and Future Directions
Uncertainty Quantification for Hallucination Detection in Large Language Models: Foundations, Methodology, and Future Directions
Sungmin Kang
Yavuz Faruk Bakman
D. Yaldiz
Baturalp Buyukates
Salman Avestimehr
HILM
72
3
0
14 Oct 2025
CPR: Mitigating Large Language Model Hallucinations with Curative Prompt Refinement
CPR: Mitigating Large Language Model Hallucinations with Curative Prompt RefinementIEEE International Conference on Systems, Man and Cybernetics (SMC), 2024
Jung-Woo Shim
Yeong-Joon Ju
Ji-Hoon Park
Seong-Whan Lee
HILMLRM
62
0
0
14 Oct 2025
Hallucination Detection via Internal States and Structured Reasoning Consistency in Large Language Models
Hallucination Detection via Internal States and Structured Reasoning Consistency in Large Language Models
Yusheng Song
Lirong Qiu
Xi Zhang
Zhihao Tang
HILMLRM
46
0
0
13 Oct 2025
The Curious Case of Factual (Mis)Alignment between LLMs' Short- and Long-Form Answers
The Curious Case of Factual (Mis)Alignment between LLMs' Short- and Long-Form Answers
Saad Obaid ul Islam
Anne Lauscher
Goran Glavaš
HILM
66
0
0
13 Oct 2025
Bolster Hallucination Detection via Prompt-Guided Data Augmentation
Bolster Hallucination Detection via Prompt-Guided Data Augmentation
Wenyun Li
Zheng Zhang
Dongmei Jiang
Xiangyuan Lan
HILM
80
0
0
13 Oct 2025
FaStfact: Faster, Stronger Long-Form Factuality Evaluations in LLMs
FaStfact: Faster, Stronger Long-Form Factuality Evaluations in LLMs
Yingjia Wan
Haochen Tan
Xiao Zhu
Xinyu Zhou
Z. Li
...
Jiaqi Zeng
Yi Xu
Jianqiao Lu
Yinhong Liu
Zhijiang Guo
HILMOffRL
167
0
0
13 Oct 2025
Discrepancy Detection at the Data Level: Toward Consistent Multilingual Question Answering
Discrepancy Detection at the Data Level: Toward Consistent Multilingual Question Answering
Lorena Calvo-Bartolomé
Valérie Aldana
Karla Cantarero
Alonso Madroñal de Mesa
Jerónimo Arenas-García
Jordan L. Boyd-Graber
HILM
92
0
0
13 Oct 2025
FactAppeal: Identifying Epistemic Factual Appeals in News Media
FactAppeal: Identifying Epistemic Factual Appeals in News Media
Guy Mor-Lan
Tamir Sheafer
Shaul R. Shenhav
HILM
68
0
0
12 Oct 2025
Detecting Hallucinations in Authentic LLM-Human Interactions
Detecting Hallucinations in Authentic LLM-Human Interactions
Yujie Ren
Niklas Gruhlke
Anne Lauscher
HILM
73
0
0
12 Oct 2025
ConsistencyAI: A Benchmark to Assess LLMs' Factual Consistency When Responding to Different Demographic Groups
ConsistencyAI: A Benchmark to Assess LLMs' Factual Consistency When Responding to Different Demographic Groups
Peter Banyas
Shristi Sharma
Alistair Simmons
Atharva Vispute
HILM
78
0
0
11 Oct 2025
On the Entity-Level Alignment in Crosslingual Consistency
On the Entity-Level Alignment in Crosslingual Consistency
Yihong Liu
Mingyang Wang
François Yvon
Hinrich Schütze
HILM
44
0
0
11 Oct 2025
Large Language Model Sourcing: A Survey
Large Language Model Sourcing: A Survey
Liang Pang
Kangxi Wu
Sunhao Dai
Zihao Wei
Zenghao Duan
...
Xiang Li
Zhiyi Yin
Jun Xu
Huawei Shen
Xueqi Cheng
HILM
47
0
0
11 Oct 2025
Large Language Models Do NOT Really Know What They Don't Know
Large Language Models Do NOT Really Know What They Don't Know
C. Cheang
Hou Pong Chan
Wenxuan Zhang
Yang Deng
HILM
55
0
0
10 Oct 2025
Enhancing Faithfulness in Abstractive Summarization via Span-Level Fine-Tuning
Enhancing Faithfulness in Abstractive Summarization via Span-Level Fine-Tuning
Sicong Huang
Qianqi Yan
Shengze Wang
Ian Lane
HILM
69
0
0
10 Oct 2025
Revisiting Hallucination Detection with Effective Rank-based Uncertainty
Revisiting Hallucination Detection with Effective Rank-based Uncertainty
Rui Wang
Zeming Wei
Guanzhang Yue
Meng Sun
UQCVHILM
125
0
0
09 Oct 2025
The Unintended Trade-off of AI Alignment:Balancing Hallucination Mitigation and Safety in LLMs
The Unintended Trade-off of AI Alignment:Balancing Hallucination Mitigation and Safety in LLMs
Omar Mahmoud
Ali Khalil
B. L. Semage
Thommen George Karimpanal
Santu Rana
HILM
48
0
0
09 Oct 2025
Loading #Papers per Month with "HILM"
Past speakers
Name (-)
Top Contributors
Name (-)
Top Organizations at ResearchTrend.AI
Name (-)
Social Events
DateLocationEvent
No social events available