Skip to content

Conversation

@845473182
Copy link
Contributor

@845473182 845473182 commented Nov 12, 2025

What this PR does / why we need it?

add DeepSeek-R1-W8A8 and Qwen3-235B-W8A8 configs in multi-nodes and EPLB scenario

Does this PR introduce any user-facing change?

no

How was this patch tested?

@github-actions
Copy link

👋 Hi! Thank you for contributing to the vLLM Ascend project. The following points will speed up your PR merge:‌‌

  • A PR should do only one thing, smaller PRs enable faster reviews.
  • Every PR should include unit tests and end-to-end tests ‌to ensure it works and is not broken by other future PRs.
  • Write the commit message by fulfilling the PR description to help reviewer and future developers understand.

If CI fails, you can run linting and testing checks locally according Contributing and Testing.

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request adds multi-node EPLB test configurations for DeepSeek-R1 and Qwen3-235B models. The configurations have a few issues that need to be addressed. For the DeepSeek-R1 config, the decoder_host_index is incorrect for a two-node decoder setup, and the engine_ids are not unique across different service instances. For the Qwen3-235B config, the benchmarks section is empty, which would prevent any tests from running. I've added comments with specific suggestions to fix these issues.

Signed-off-by: 白永斌 <[email protected]>
@wangxiyuan wangxiyuan merged commit f90ed95 into vllm-project:main Nov 14, 2025
23 checks passed
luolun pushed a commit to luolun/vllm-ascend that referenced this pull request Nov 19, 2025
…A8 (vllm-project#4144)

### What this PR does / why we need it?
add DeepSeek-R1-W8A8 and Qwen3-235B-W8A8 configs in multi-nodes and EPLB
scenario

### Does this PR introduce _any_ user-facing change?
no

- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: 白永斌 <[email protected]>
Co-authored-by: 白永斌 <[email protected]>
Signed-off-by: luolun <[email protected]>
hwhaokun pushed a commit to hwhaokun/vllm-ascend that referenced this pull request Nov 19, 2025
…A8 (vllm-project#4144)

### What this PR does / why we need it?
add DeepSeek-R1-W8A8 and Qwen3-235B-W8A8 configs in multi-nodes and EPLB
scenario

### Does this PR introduce _any_ user-facing change?
no

- vLLM version: v0.11.0
- vLLM main:
vllm-project/vllm@83f478b

---------

Signed-off-by: 白永斌 <[email protected]>
Co-authored-by: 白永斌 <[email protected]>
Signed-off-by: hwhaokun <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants