Skip to content

Conversation

@wangxiyuan
Copy link
Collaborator

@wangxiyuan wangxiyuan commented Nov 25, 2025

vLLM graph feature now rely on torch >=2.8. To make graph mode work, we need upgrade torch version as well. For long term support, upgrade torch to a newer one is good to go as well.

Related vLLM change: vllm-project/vllm#25110

@wangxiyuan wangxiyuan added ready read for review ready-for-test start test by label for PR labels Nov 25, 2025
@github-actions github-actions bot added the documentation Improvements or additions to documentation label Nov 25, 2025
@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request upgrades the required versions of torch and torch-npu from 2.7.1 to 2.8.0. The changes are applied consistently across all relevant files, including the CMake build configuration (CMakeLists.txt), Python dependency files (pyproject.toml, requirements.txt), and documentation (README.md, README.zh.md). The changes are correct and align with the goal of the pull request. I have not found any issues of high or critical severity.

@wangxiyuan wangxiyuan force-pushed the upgrade_torch_npu branch 2 times, most recently from b3935b7 to 708ebbc Compare November 26, 2025 06:45
@wangxiyuan wangxiyuan force-pushed the upgrade_torch_npu branch 2 times, most recently from c61e8cb to bf448d0 Compare November 27, 2025 11:10
@zhangxinyuehfad
Copy link
Contributor

torchair + v1 scheduler + torch2.8 test failed

command:

llm serve vllm-ascend/DeepSeek-V3-W8A8 \
 --no-enable-prefix-caching \
 --max-num-seqs 16 \
 --tensor-parallel-size 16 \
 --max-model-len 36864 \
 --max-num-batched-tokens 6000 \
 --enable-expert-parallel \
 --trust-remote-code \
 --quantization ascend \
 --gpu-memory-utilization 0.9 \
 --speculative-config '{"num_speculative_tokens": 1, "method":"deepseek_mtp"}' \
 --additional-config '{"ascend_scheduler_config":{"enabled":false},"torchair_graph_config":{"enabled":true,"enable_multistream_moe":true},"chunked_prefill_for_mla":true,"enable_weight_nz_layout":true}' &  
error log
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815] Traceback (most recent call last):
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch_npu/dynamo/torchair/_utils/error_code.py", line 43, in wapper
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return func(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch_npu/dynamo/torchair/core/_backend.py", line 137, in run
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return super(TorchNpuGraph, self).run((inputs, assigned_outputs, stream))
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815] RuntimeError: [PID: 1762949] 2025-11-27-12:35:42.886.865 Insufficient_Resources(EL0006): The resources are insufficient.
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Solution: Close applications not in use.
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         TraceBack (most recent call last):
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         [SqCqManage]Alloc sq cq fail, stream_id=1984, retCode=0x7020023.[FUNC:AllocStreamSqCq][FILE:stream_sqcq_manage.cc][LINE:79]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         [SqCqManage]Alloc sq cq fail, stream_id=1984, retCode=0x7020023.[FUNC:Setup][FILE:stream.cc][LINE:653]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         rtStreamCreateWithFlags execute failed, reason=[driver error:resource alloc fail][FUNC:FuncErrorReason][FILE:error_message_manage.cc][LINE:53]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Assert ((rtStreamCreateWithFlags(&stream, priority, stream_flag)) == 0) failed[FUNC:CreateNewStream][FILE:reusable_stream_allocator.cc][LINE:72]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Assert ((stream_status) != nullptr) failed[FUNC:GetOrCreateRtStream][FILE:reusable_stream_allocator.cc][LINE:88]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Assert ((reusable_stream_allocator_->GetOrCreateRtStream(stream, runtime_model_id_, priority_, stream_flags, task_num)) == ge::SUCCESS) failed[FUNC:InitRuntimeResource][FILE:davinci_model.cc][LINE:1312]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         davinci model init failed[FUNC:DavinciModelCreate][FILE:davinci_model_kernel.cc][LINE:249]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Failed to execute init graph[FUNC:Load][FILE:model_v2_executor.cc][LINE:189]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Failed load rt v2 model for graph graph_1[FUNC:Init][FILE:hybrid_model_rt_v2_executor.cc][LINE:826]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Load graph failed, error code:1343225857, session_id:6, graph_id:0, isEnableSliceSchedule:0.[FUNC:LoadGraph][FILE:ge_api.cc][LINE:994]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815] During handling of the above exception, another exception occurred:
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815] Traceback (most recent call last):
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm/vllm/v1/executor/multiproc_executor.py", line 810, in worker_busy_loop
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     output = func(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]              ^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/worker/worker_v1.py", line 349, in compile_or_warm_up_model
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     self.model_runner.capture_model()
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/worker/model_runner_v1.py", line 4042, in capture_model
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     self._capture_model()
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/torchair/torchair_model_runner.py", line 302, in _capture_model
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     self._compile_torchair_graph(torchair_graph_batch_sizes)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/torchair/torchair_model_runner.py", line 261, in _compile_torchair_graph
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     self._dummy_run(num_tokens, is_torchair_compile=True)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return func(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/worker/model_runner_v1.py", line 3032, in _dummy_run
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     hidden_states = self._generate_dummy_run_hidden_states(
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]                     ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/torchair/torchair_model_runner.py", line 239, in _generate_dummy_run_hidden_states
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     hidden_states = compiled_model(
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]                     ^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 375, in __call__
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return super().__call__(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return self._call_impl(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return forward_call(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 736, in compile_wrapper
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return fn(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return self._call_impl(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return forward_call(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/vllm-workspace/vllm-ascend/vllm_ascend/torchair/models/torchair_deepseek_v2.py", line 1335, in forward
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     def forward(
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_dynamo/eval_frame.py", line 929, in _fn
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return fn(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_functorch/aot_autograd.py", line 1241, in forward
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return compiled_fn(full_args)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 384, in runtime_wrapper
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     all_outs = call_func_at_runtime_with_args(
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_functorch/_aot_autograd/utils.py", line 126, in call_func_at_runtime_with_args
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     out = normalize_as_list(f(args))
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]                             ^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 750, in inner_fn
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     outs = compiled_fn(args)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_functorch/_aot_autograd/runtime_wrappers.py", line 556, in wrapper
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return compiled_fn(runtime_args)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch/_functorch/_aot_autograd/utils.py", line 100, in g
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return f(*args)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch_npu/dynamo/torchair/npu_fx_compiler.py", line 355, in __call__
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     gm_result = self.runner(*args, **kwargs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch_npu/dynamo/torchair/_ge_concrete_graph/fx2ge_converter.py", line 692, in __call__
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     ge_outputs = self.graph.run(inputs)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]                  ^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch_npu/dynamo/torchair/ge/_ge_graph.py", line 674, in run
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     return self._executor.run(inputs, assigned_outputs, stream)
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]            ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]   File "/usr/local/python3.11.13/lib/python3.11/site-packages/torch_npu/dynamo/torchair/_utils/error_code.py", line 46, in wapper
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]     raise type(e)("\n".join(msg))
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815] RuntimeError: [PID: 1762949] 2025-11-27-12:35:42.886.865 Insufficient_Resources(EL0006): The resources are insufficient.
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Solution: Close applications not in use.
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         TraceBack (most recent call last):
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         [SqCqManage]Alloc sq cq fail, stream_id=1984, retCode=0x7020023.[FUNC:AllocStreamSqCq][FILE:stream_sqcq_manage.cc][LINE:79]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         [SqCqManage]Alloc sq cq fail, stream_id=1984, retCode=0x7020023.[FUNC:Setup][FILE:stream.cc][LINE:653]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         rtStreamCreateWithFlags execute failed, reason=[driver error:resource alloc fail][FUNC:FuncErrorReason][FILE:error_message_manage.cc][LINE:53]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Assert ((rtStreamCreateWithFlags(&stream, priority, stream_flag)) == 0) failed[FUNC:CreateNewStream][FILE:reusable_stream_allocator.cc][LINE:72]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Assert ((stream_status) != nullptr) failed[FUNC:GetOrCreateRtStream][FILE:reusable_stream_allocator.cc][LINE:88]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Assert ((reusable_stream_allocator_->GetOrCreateRtStream(stream, runtime_model_id_, priority_, stream_flags, task_num)) == ge::SUCCESS) failed[FUNC:InitRuntimeResource][FILE:davinci_model.cc][LINE:1312]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         davinci model init failed[FUNC:DavinciModelCreate][FILE:davinci_model_kernel.cc][LINE:249]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Failed to execute init graph[FUNC:Load][FILE:model_v2_executor.cc][LINE:189]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Failed load rt v2 model for graph graph_1[FUNC:Init][FILE:hybrid_model_rt_v2_executor.cc][LINE:826]
(Worker_TP2_EP2 pid=1762949) ERROR 11-27 12:35:45 [multiproc_executor.py:815]         Load graph failed, error code:1343225857, session_id:6, graph_id:0, isEnableSliceSchedule:0.[FUNC:LoadGraph][FILE:ge_api.cc][LINE:994]

@wangxiyuan wangxiyuan force-pushed the upgrade_torch_npu branch 2 times, most recently from 300d5c4 to 7024251 Compare November 28, 2025 02:13
@github-actions
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@github-actions
Copy link

This pull request has conflicts, please resolve those before we can evaluate the pull request.

@wangxiyuan wangxiyuan force-pushed the upgrade_torch_npu branch 2 times, most recently from ebbdffa to acfac40 Compare November 28, 2025 15:24
Signed-off-by: wangxiyuan <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

documentation Improvements or additions to documentation module:core module:quantization module:tests ready read for review ready-for-test start test by label for PR

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants