...

lihanc111

75

Karma

2025-06-18

Created

Recent Activity

  • Please send to contact@lmcache.ai

  • It is almost true for both. Although for the second case you can just skip storing in these cases where there is little improvement.

  • It is in IBM's llm-d open source stack

  • Our team has built this open source project, LMCache, to reduce repetitive computation in LLM inference and make systems serve more people (3x more throughput in chat applications) and it has been used in IBM's open source LLM inference stack.

    In LLM serving, the input is computed into intermediate states called KV cache to further provide answers. These data are relatively large (~1-2GB for long context) and are often evicted when GPU memory is not enough. In these cases, when users ask a follow up question, the software needs to recompute for the same KV Cache. LMCache is designed to combat that by efficiently offloading and loading these KV cache to and from DRAM and disk.

    Ask us anything!

HackerNews