Assessing LLM Reasoning Steps via Principal Knowledge GroundingConference on Empirical Methods in Natural Language Processing (EMNLP), 2025 |
AttnCache: Accelerating Self-Attention Inference for LLM Prefill via Attention CacheIACR Cryptology ePrint Archive (IACR ePrint), 2025 |