-
Notifications
You must be signed in to change notification settings - Fork 613
upgrade torch npu version #4433
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
|
👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:
If CI fails, you can run linting and testing checks locally according Contributing and Testing. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request upgrades the required versions of torch and torch-npu from 2.7.1 to 2.8.0. The changes are applied consistently across all relevant files, including the CMake build configuration (CMakeLists.txt), Python dependency files (pyproject.toml, requirements.txt), and documentation (README.md, README.zh.md). The changes are correct and align with the goal of the pull request. I have not found any issues of high or critical severity.
b3935b7 to
708ebbc
Compare
708ebbc to
9d131fa
Compare
c61e8cb to
bf448d0
Compare
|
torchair + v1 scheduler + torch2.8 test failed command: llm serve vllm-ascend/DeepSeek-V3-W8A8 \
--no-enable-prefix-caching \
--max-num-seqs 16 \
--tensor-parallel-size 16 \
--max-model-len 36864 \
--max-num-batched-tokens 6000 \
--enable-expert-parallel \
--trust-remote-code \
--quantization ascend \
--gpu-memory-utilization 0.9 \
--speculative-config '{"num_speculative_tokens": 1, "method":"deepseek_mtp"}' \
--additional-config '{"ascend_scheduler_config":{"enabled":false},"torchair_graph_config":{"enabled":true,"enable_multistream_moe":true},"chunked_prefill_for_mla":true,"enable_weight_nz_layout":true}' & error log |
300d5c4 to
7024251
Compare
7024251 to
9b88dd7
Compare
|
This pull request has conflicts, please resolve those before we can evaluate the pull request. |
4d23315 to
9186dc2
Compare
|
This pull request has conflicts, please resolve those before we can evaluate the pull request. |
ebbdffa to
acfac40
Compare
Signed-off-by: wangxiyuan <[email protected]>
acfac40 to
5dc99cf
Compare
vLLM graph feature now rely on torch >=2.8. To make graph mode work, we need upgrade torch version as well. For long term support, upgrade torch to a newer one is good to go as well.
Related vLLM change: vllm-project/vllm#25110