ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2312.17484
  4. Cited By
Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models
  through Intervention without Tuning
v1v2 (latest)

Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning

AAAI Conference on Artificial Intelligence (AAAI), 2023
29 December 2023
Zhongzhi Chen
Xingwu Sun
Xianfeng Jiao
Fengzong Lian
Zhanhui Kang
Di Wang
Cheng-zhong Xu
    HILM
ArXiv (abs)PDFHTML

Papers citing "Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning"

27 / 27 papers shown
TraceDet: Hallucination Detection from the Decoding Trace of Diffusion Large Language Models
TraceDet: Hallucination Detection from the Decoding Trace of Diffusion Large Language Models
Shenxu Chang
Junchi Yu
Weixing Wang
Yongqiang Chen
Jialin Yu
Philip Torr
Jindong Gu
HILM
156
0
0
30 Sep 2025
Steering When Necessary: Flexible Steering Large Language Models with Backtracking
Steering When Necessary: Flexible Steering Large Language Models with Backtracking
Jinwei Gan
Zifeng Cheng
Zhiwei Jiang
Cong Wang
Yafeng Yin
Xiang Luo
Yuchen Fu
Qing Gu
KELMLLMSV
187
1
0
25 Aug 2025
Expanding before Inferring: Enhancing Factuality in Large Language Models through Premature Layers Interpolation
Expanding before Inferring: Enhancing Factuality in Large Language Models through Premature Layers Interpolation
Dingwei Chen
Ziqiang Liu
Feiteng Fang
Chak Tou Leong
Shiwen Ni
A. Argha
Hamid Alinejad-Rokny
Min Yang
Chengming Li
KELMHILM
314
2
0
03 Jun 2025
ExpertSteer: Intervening in LLMs through Expert Knowledge
ExpertSteer: Intervening in LLMs through Expert Knowledge
Weixuan Wang
Minghao Wu
Barry Haddow
Alexandra Birch
LLMSV
494
1
0
18 May 2025
The Illusionist's Prompt: Exposing the Factual Vulnerabilities of Large Language Models with Linguistic Nuances
The Illusionist's Prompt: Exposing the Factual Vulnerabilities of Large Language Models with Linguistic Nuances
Yining Wang
Longji Xu
Xi Li
Mi Zhang
Geng Hong
Min Yang
AAMLHILM
240
1
0
01 Apr 2025
Personalized Text Generation with Contrastive Activation SteeringAnnual Meeting of the Association for Computational Linguistics (ACL), 2025
Jinghao Zhang
Yi Liu
Wenjie Wang
Sihan Yang
Shu Wu
Liang Wang
Tat-Seng Chua
LLMSV
267
14
0
07 Mar 2025
DSVD: Dynamic Self-Verify Decoding for Faithful Generation in Large Language Models
Y. Guo
Yuchen Yang
Zhe Chen
Pingjie Wang
Yusheng Liao
Yujiao Shi
Yanfeng Wang
Yu Wang
HILM
301
2
0
05 Mar 2025
SAFE: A Sparse Autoencoder-Based Framework for Robust Query Enrichment and Hallucination Mitigation in LLMs
Samir Abdaljalil
Filippo Pallucchini
Andrea Seveso
Hasan Kurban
Fabio Mercorio
Erchin Serpedin
HILM
264
6
0
04 Mar 2025
Steer LLM Latents for Hallucination Detection
Steer LLM Latents for Hallucination Detection
Seongheon Park
Xuefeng Du
Min-Hsuan Yeh
Haobo Wang
Yixuan Li
LLMSV
317
3
0
01 Mar 2025
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful Comparators
Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful ComparatorsAAAI Conference on Artificial Intelligence (AAAI), 2024
Jinjie Wei
Dongling Xiao
Jinjie Wei
Mingcheng Li
Zhaoyu Chen
Ke Li
Li Zhang
HILM
515
15
0
28 Jan 2025
Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large
  Vision-Language Model via Causality Analysis
Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality AnalysisIEEE Workshop/Winter Conference on Applications of Computer Vision (WACV), 2024
Po-Hsuan Huang
Jeng-Lin Li
Chin-Po Chen
Ming-Ching Chang
Wei-Chao Chen
LRM
301
4
0
04 Dec 2024
Distinguishing Ignorance from Error in LLM Hallucinations
Distinguishing Ignorance from Error in LLM Hallucinations
Adi Simhi
Jonathan Herzig
Idan Szpektor
Yonatan Belinkov
HILM
216
13
0
29 Oct 2024
Semantics-Adaptive Activation Intervention for LLMs via Dynamic Steering Vectors
Semantics-Adaptive Activation Intervention for LLMs via Dynamic Steering VectorsInternational Conference on Learning Representations (ICLR), 2024
Weixuan Wang
J. Yang
Wei Peng
LLMSV
327
23
0
16 Oct 2024
NoVo: Norm Voting off Hallucinations with Attention Heads in Large
  Language Models
NoVo: Norm Voting off Hallucinations with Attention Heads in Large Language ModelsInternational Conference on Learning Representations (ICLR), 2024
Zheng Yi Ho
Yaning Tan
Sen Zhang
Yibing Zhan
Dacheng Tao
304
5
0
11 Oct 2024
Lower Layers Matter: Alleviating Hallucination via Multi-Layer Fusion Contrastive Decoding with Truthfulness Refocused
Lower Layers Matter: Alleviating Hallucination via Multi-Layer Fusion Contrastive Decoding with Truthfulness Refocused
Dingwei Chen
Feiteng Fang
Shiwen Ni
Feng Liang
Xiping Hu
A. Argha
Hamid Alinejad-Rokny
Min Yang
Chengming Li
HILM
246
3
0
16 Aug 2024
Internal Consistency and Self-Feedback in Large Language Models: A
  Survey
Internal Consistency and Self-Feedback in Large Language Models: A Survey
Xun Liang
Chenyang Xi
Zifan Zheng
Ding Chen
Qingchen Yu
...
Rong-Hua Li
Peng Cheng
Zhonghao Wang
Feiyu Xiong
Zhiyu Li
HILMLRM
497
45
0
19 Jul 2024
Lookback Lens: Detecting and Mitigating Contextual Hallucinations in
  Large Language Models Using Only Attention Maps
Lookback Lens: Detecting and Mitigating Contextual Hallucinations in Large Language Models Using Only Attention Maps
Yung-Sung Chuang
Linlu Qiu
Cheng-Yu Hsieh
Ranjay Krishna
Yoon Kim
James R. Glass
HILM
247
84
0
09 Jul 2024
Mitigating Large Language Model Hallucination with Faithful Finetuning
Mitigating Large Language Model Hallucination with Faithful Finetuning
Minda Hu
Bowei He
Yufei Wang
Liangyou Li
Chen Ma
Irwin King
HILM
285
22
0
17 Jun 2024
PrivacyRestore: Privacy-Preserving Inference in Large Language Models via Privacy Removal and Restoration
PrivacyRestore: Privacy-Preserving Inference in Large Language Models via Privacy Removal and Restoration
Huiping Zhuang
Jianwei Wang
Zhengdong Lu
Huiping Zhuang
Haoran Li
Huiping Zhuang
Cen Chen
RALMKELM
661
17
0
03 Jun 2024
Spectral Editing of Activations for Large Language Model Alignment
Spectral Editing of Activations for Large Language Model AlignmentNeural Information Processing Systems (NeurIPS), 2024
Yifu Qiu
Zheng Zhao
Yftah Ziser
Anna Korhonen
Edoardo Ponti
Shay B. Cohen
KELMLLMSV
400
40
0
15 May 2024
Enhanced Language Model Truthfulness with Learnable Intervention and
  Uncertainty Expression
Enhanced Language Model Truthfulness with Learnable Intervention and Uncertainty Expression
Farima Fatahi Bayat
Xin Liu
H. V. Jagadish
Lu Wang
HILMKELM
260
2
0
01 May 2024
Constructing Benchmarks and Interventions for Combating Hallucinations
  in LLMs
Constructing Benchmarks and Interventions for Combating Hallucinations in LLMs
Adi Simhi
Jonathan Herzig
Idan Szpektor
Yonatan Belinkov
HILM
287
20
0
15 Apr 2024
Non-Linear Inference Time Intervention: Improving LLM Truthfulness
Non-Linear Inference Time Intervention: Improving LLM Truthfulness
Jakub Hoscilowicz
Adam Wiacek
Jan Chojnacki
Adam Cieślak
Leszek Michon
Vitalii Urbanevych
Artur Janicki
KELM
167
5
0
27 Mar 2024
HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild
HaluEval-Wild: Evaluating Hallucinations of Language Models in the Wild
Zhiying Zhu
Yiming Yang
Zhiqing Sun
HILMVLM
281
25
0
07 Mar 2024
TruthX: Alleviating Hallucinations by Editing Large Language Models in
  Truthful Space
TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful Space
Shaolei Zhang
Tian Yu
Yang Feng
HILMKELM
313
81
0
27 Feb 2024
GRATH: Gradual Self-Truthifying for Large Language Models
GRATH: Gradual Self-Truthifying for Large Language ModelsInternational Conference on Machine Learning (ICML), 2024
Weixin Chen
Basel Alomair
Yue Liu
HILMSyDa
126
7
0
22 Jan 2024
Zero-Resource Hallucination Prevention for Large Language Models
Zero-Resource Hallucination Prevention for Large Language ModelsConference on Empirical Methods in Natural Language Processing (EMNLP), 2023
Junyu Luo
Cao Xiao
Fenglong Ma
HILM
489
35
0
06 Sep 2023
1