Communities
Connect sessions
AI calendar
Organizations
Join Slack
Contact Sales
Search
Open menu
Home
Papers
2508.04462
Cited By
v1
v2 (latest)
CARD: A Cache-Assisted Parallel Speculative Decoding Framework via Query-and-Correct Paradigm for Accelerating LLM Inference
6 August 2025
Enyu Zhou
Kai Sheng
Hao Chen
Xin He
LRM
Re-assign community
ArXiv (abs)
PDF
HTML
Github (2★)
Papers citing
"CARD: A Cache-Assisted Parallel Speculative Decoding Framework via Query-and-Correct Paradigm for Accelerating LLM Inference"
0 / 0 papers shown
No papers found