Skip to content

Releases: LMCache/lmcache-vllm

v0.6.2.3

10 Dec 20:05
f5b015b
Compare
Choose a tag to compare
v0.6.2.3 Pre-release
Pre-release

What's Changed

New Contributors

Full Changelog: v0.6.2.2...v0.6.2.3

v0.6.2.2

29 Oct 23:12
b890544
Compare
Choose a tag to compare
v0.6.2.2 Pre-release
Pre-release

New version: v0.6.2.2

Compatibility:

  • vLLM: 0.6.1.post2, 0.6.2
  • LMCache: 0.1.3

Key Features

  • Supporting chunked prefill in vLLM
  • Faster KV loading for multi-turn conversation by saving KV at the decoding time
  • Experimental KV blending feature to enable reusing non-prefix KV caches
  • New model support: llama-3.1 and qwen-2

What's Changed

New Contributors

Full Changelog: v0.6.2.post1...v0.6.2.2

v0.6.2.post1

24 Sep 19:56
f2bc279
Compare
Choose a tag to compare
v0.6.2.post1 Pre-release
Pre-release

What's Changed

  • [Release] lmcache_vllm wrapper with vllm 0.6.1.post2 support by @ApostaC in #9
  • Fix the main-thread-blocker in lmcache_kv_store

v0.1.1-alpha

17 Sep 18:43
bbbfb39
Compare
Choose a tag to compare
v0.1.1-alpha Pre-release
Pre-release

Tagging with LMCache v0.1.1-alpha