ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.14852
  4. Cited By
Cost-Efficient Serving of LLM Agents via Test-Time Plan Caching

Cost-Efficient Serving of LLM Agents via Test-Time Plan Caching

17 June 2025
Qizheng Zhang
Michael Wornow
Kunle Olukotun
ArXiv (abs)PDFHTML

Papers citing "Cost-Efficient Serving of LLM Agents via Test-Time Plan Caching"

2 / 2 papers shown
Continuum: Efficient and Robust Multi-Turn LLM Agent Scheduling with KV Cache Time-to-Live
Continuum: Efficient and Robust Multi-Turn LLM Agent Scheduling with KV Cache Time-to-Live
Hanchen Li
Qiuyang Mang
Runyuan He
Qizheng Zhang
Huanzhi Mao
Xiaokun Chen
Alvin Cheung
Joseph E. Gonzalez
Ion Stoica
Ion Stoica
196
2
0
04 Nov 2025
Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models
Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models
Qizheng Zhang
Changran Hu
Shubhangi Upasani
Boyuan Ma
Fenglu Hong
...
Mengmeng Ji
Hanchen Li
Urmish Thakker
James Zou
Kunle Olukotun
LLMAGKELM
223
29
0
06 Oct 2025
1
Page 1 of 1