Skip to content

LMDeploy Release V0.0.8

Compare
Choose a tag to compare
@lvhan028 lvhan028 released this 11 Sep 15:34
· 935 commits to main since this release
450757b

Highlight

  • Support Baichuan2-7B-Base and Baichuan2-7B-Chat
  • Support all features of Code Llama: code completion, infilling, chat / instruct, and python specialist

What's Changed

🚀 Features

🐞 Bug fixes

  • [Fix] when using stream is False, continuous batching doesn't work by @sleepwalker2017 in #346
  • [Fix] Set max dynamic smem size for decoder MHA to support context length > 8k by @lvhan028 in #377
  • Fix exceed session len core dump for chat and generate by @AllentDan in #366
  • [Fix] update puyu model by @Harold-lkk in #399

📚 Documentations

New Contributors

Full Changelog: v0.0.7...v0.0.8