Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2312.17484
Cited By
v1
v2 (latest)
Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning
AAAI Conference on Artificial Intelligence (AAAI), 2023
29 December 2023
Zhongzhi Chen
Xingwu Sun
Xianfeng Jiao
Fengzong Lian
Zhanhui Kang
Di Wang
Cheng-zhong Xu
HILM
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning"
27 / 27 papers shown
TraceDet: Hallucination Detection from the Decoding Trace of Diffusion Large Language Models
Shenxu Chang
Junchi Yu
Weixing Wang
Yongqiang Chen
Jialin Yu
Philip Torr
Jindong Gu
HILM
156
0
0
30 Sep 2025
Steering When Necessary: Flexible Steering Large Language Models with Backtracking
Jinwei Gan
Zifeng Cheng
Zhiwei Jiang
Cong Wang
Yafeng Yin
Xiang Luo
Yuchen Fu
Qing Gu
KELM
LLMSV
187
1
0
25 Aug 2025
Expanding before Inferring: Enhancing Factuality in Large Language Models through Premature Layers Interpolation
Dingwei Chen
Ziqiang Liu
Feiteng Fang
Chak Tou Leong
Shiwen Ni
A. Argha
Hamid Alinejad-Rokny
Min Yang
Chengming Li
KELM
HILM
314
2
0
03 Jun 2025
ExpertSteer: Intervening in LLMs through Expert Knowledge
Weixuan Wang
Minghao Wu
Barry Haddow
Alexandra Birch
LLMSV
494
1
0
18 May 2025
The Illusionist's Prompt: Exposing the Factual Vulnerabilities of Large Language Models with Linguistic Nuances
Yining Wang
Longji Xu
Xi Li
Mi Zhang
Geng Hong
Min Yang
AAML
HILM
240
1
0
01 Apr 2025
Personalized Text Generation with Contrastive Activation Steering
Annual Meeting of the Association for Computational Linguistics (ACL), 2025
Jinghao Zhang
Yi Liu
Wenjie Wang
Sihan Yang
Shu Wu
Liang Wang
Tat-Seng Chua
LLMSV
267
14
0
07 Mar 2025
DSVD: Dynamic Self-Verify Decoding for Faithful Generation in Large Language Models
Y. Guo
Yuchen Yang
Zhe Chen
Pingjie Wang
Yusheng Liao
Yujiao Shi
Yanfeng Wang
Yu Wang
HILM
301
2
0
05 Mar 2025
SAFE: A Sparse Autoencoder-Based Framework for Robust Query Enrichment and Hallucination Mitigation in LLMs
Samir Abdaljalil
Filippo Pallucchini
Andrea Seveso
Hasan Kurban
Fabio Mercorio
Erchin Serpedin
HILM
264
6
0
04 Mar 2025
Steer LLM Latents for Hallucination Detection
Seongheon Park
Xuefeng Du
Min-Hsuan Yeh
Haobo Wang
Yixuan Li
LLMSV
317
3
0
01 Mar 2025
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful Comparators
AAAI Conference on Artificial Intelligence (AAAI), 2024
Jinjie Wei
Dongling Xiao
Jinjie Wei
Mingcheng Li
Zhaoyu Chen
Ke Li
Li Zhang
HILM
515
15
0
28 Jan 2025
Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality Analysis
IEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2024
Po-Hsuan Huang
Jeng-Lin Li
Chin-Po Chen
Ming-Ching Chang
Wei-Chao Chen
LRM
301
4
0
04 Dec 2024
Distinguishing Ignorance from Error in LLM Hallucinations
Adi Simhi
Jonathan Herzig
Idan Szpektor
Yonatan Belinkov
HILM
216
13
0
29 Oct 2024
Semantics-Adaptive Activation Intervention for LLMs via Dynamic Steering Vectors
International Conference on Learning Representations (ICLR), 2024
Weixuan Wang
J. Yang
Wei Peng
LLMSV
327
23
0
16 Oct 2024
NoVo: Norm Voting off Hallucinations with Attention Heads in Large Language Models
International Conference on Learning Representations (ICLR), 2024
Zheng Yi Ho
Yaning Tan
Sen Zhang
Yibing Zhan
Dacheng Tao
304
5
0
11 Oct 2024
Lower Layers Matter: Alleviating Hallucination via Multi-Layer Fusion Contrastive Decoding with Truthfulness Refocused
Dingwei Chen
Feiteng Fang
Shiwen Ni
Feng Liang
Xiping Hu
A. Argha
Hamid Alinejad-Rokny
Min Yang
Chengming Li
HILM
246
3
0
16 Aug 2024
Internal Consistency and Self-Feedback in Large Language Models: A Survey
Xun Liang
Chenyang Xi
Zifan Zheng
Ding Chen
Qingchen Yu
...
Rong-Hua Li
Peng Cheng
Zhonghao Wang
Feiyu Xiong
Zhiyu Li
HILM
LRM
497
45
0
19 Jul 2024
Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
Yung-Sung Chuang
Linlu Qiu
Cheng-Yu Hsieh
Ranjay Krishna
Yoon Kim
James R. Glass
HILM
247
84
0
09 Jul 2024
Mitigating Large Language Model Hallucination with Faithful Finetuning
Minda Hu
Bowei He
Yufei Wang
Liangyou Li
Chen Ma
Irwin King
HILM
285
22
0
17 Jun 2024
PrivacyRestore: Privacy-Preserving Inference in Large Language Models via Privacy Removal and Restoration
Huiping Zhuang
Jianwei Wang
Zhengdong Lu
Huiping Zhuang
Haoran Li
Huiping Zhuang
Cen Chen
RALM
KELM
661
17
0
03 Jun 2024
Spectral Editing of Activations for Large Language Model Alignment
Neural Information Processing Systems (NeurIPS), 2024
Yifu Qiu
Zheng Zhao
Yftah Ziser
Anna Korhonen
Edoardo Ponti
Shay B. Cohen
KELM
LLMSV
400
40
0
15 May 2024
Enhanced Language Model Truthfulness with Learnable Intervention and Uncertainty Expression
Farima Fatahi Bayat
Xin Liu
H. V. Jagadish
Lu Wang
HILM
KELM
260
2
0
01 May 2024
Constructing Benchmarks and Interventions for Combating Hallucinations in LLMs
Adi Simhi
Jonathan Herzig
Idan Szpektor
Yonatan Belinkov
HILM
287
20
0
15 Apr 2024
Non-Linear Inference Time Intervention: Improving LLM Truthfulness
Jakub Hoscilowicz
Adam Wiacek
Jan Chojnacki
Adam Cieślak
Leszek Michon
Vitalii Urbanevych
Artur Janicki
KELM
167
5
0
27 Mar 2024
HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild
Zhiying Zhu
Yiming Yang
Zhiqing Sun
HILM
VLM
281
25
0
07 Mar 2024
TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space
Shaolei Zhang
Tian Yu
Yang Feng
HILM
KELM
313
81
0
27 Feb 2024
GRATH: Gradual Self-Truthifying for Large Language Models
International Conference on Machine Learning (ICML), 2024
Weixin Chen
Basel Alomair
Yue Liu
HILM
SyDa
126
7
0
22 Jan 2024
Zero-Resource Hallucination Prevention for Large Language Models
Conference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Junyu Luo
Cao Xiao
Fenglong Ma
HILM
489
35
0
06 Sep 2023
1