LMDeploy Release V0.0.8
Highlight
- Support Baichuan2-7B-Base and Baichuan2-7B-Chat
- Support all features of Code Llama: code completion, infilling, chat / instruct, and python specialist
What's Changed
🚀 Features
- Support baichuan2-chat chat template by @wangruohui in #378
- Support codellama by @lvhan028 in #359
🐞 Bug fixes
- [Fix] when using stream is False, continuous batching doesn't work by @sleepwalker2017 in #346
- [Fix] Set max dynamic smem size for decoder MHA to support context length > 8k by @lvhan028 in #377
- Fix exceed session len core dump for chat and generate by @AllentDan in #366
- [Fix] update puyu model by @Harold-lkk in #399
📚 Documentations
- [Docs] Fix quantization docs link by @LZHgrla in #367
- [Docs] Simplify
build.md
by @pppppM in #370 - [Docs] Update lmdeploy logo by @lvhan028 in #372
New Contributors
- @sleepwalker2017 made their first contribution in #346
Full Changelog: v0.0.7...v0.0.8