LMDeploy Release V0.0.10
What's Changed
💥 Improvements
- [feature] Graceful termination of background threads in LlamaV2 by @akhoroshev in #458
- expose stop words and filter eoa by @AllentDan in #352
🐞 Bug fixes
- Fix side effect brought by supporting codellama:
sequence_start
is always true when callingmodel.get_prompt
by @lvhan028 in #466 - Miss meta instruction of internlm-chat model by @lvhan028 in #470
- [bug] Fix race condition by @akhoroshev in #460
- Fix compatibility issues with Pydantic 2 by @aisensiy in #465
- fix benchmark serving cannot use Qwen tokenizer by @AllentDan in #443
- Fix memory leak by @lvhan028 in #488
📚 Documentations
- Fix typo in README.md by @eltociear in #462
🌐 Other
New Contributors
- @eltociear made their first contribution in #462
- @akhoroshev made their first contribution in #458
- @aisensiy made their first contribution in #465
Full Changelog: v0.0.9...v0.0.10