161
v1v2 (latest)

PruneCD: Contrasting Pruned Self Model to Improve Decoding Factuality

Main:4 Pages
10 Figures
Bibliography:2 Pages
10 Tables
Appendix:6 Pages
Abstract

To mitigate the hallucination problem in large language models, DoLa exploits early exit logits from the same model as a contrastive prior. However, we found that these early exit logits tend to be flat, low in magnitude, and fail to reflect meaningful contrasts. To address this, we propose PruneCD, a novel contrastive decoding method that constructs the amateur model via layer pruning rather than early exit. This design leads to more informative and well-aligned logits, enabling more effective contrastive decoding. Through qualitative and quantitative analyses, we demonstrate that PruneCD consistently improves factuality with minimal inference overhead, offering a robust and practical approach to mitigating hallucinations in LLMs.

View on arXiv
Comments on this paper