Skip to content

Conversation

@wangxiyuan
Copy link
Collaborator

@wangxiyuan wangxiyuan commented Dec 17, 2025

Pick from #5098 to fix the doc error

MengqingCao and others added 2 commits December 17, 2025 09:00
Signed-off-by: wangxiyuan <[email protected]>
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request updates documentation to align with the v0.11.0 release, including version numbers, feature statuses, and links. My review identified a couple of incorrect links in the feature support matrix which could negatively impact users. The feedback provided aims to correct these documentation errors.

| Guided Decoding | 🟢 Functional | See detailed note: [Structured Output Guide][guided_decoding] |
| Tensor Parallel | 🟢 Functional | Make TP >4 work with graph mode. |
| Pipeline Parallel | 🟡 Planned | Broken in this version, will fix in next release. |
| Expert Parallel | 🟢 Functional | See detailed note: [Expert Load Balance (EPLB)][graph_mode] |
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

The link for "Expert Parallel" points to the "Graph Mode Guide", which does not seem to contain information about Expert Load Balance (EPLB). The relevant configuration details for EPLB are in the "Additional Configuration" guide. Please update the link to point to the correct document.

For example, you could change the reference to [eplb] and add a new link definition for it at the end of the file:
[eplb]: https://docs.vllm.ai/projects/ascend/en/v0.11.0-dev/user_guide/configuration/additional_config.html

Suggested change
| Expert Parallel | 🟢 Functional | See detailed note: [Expert Load Balance (EPLB)][graph_mode] |
| Expert Parallel | 🟢 Functional | See detailed note: [Expert Load Balance (EPLB)][eplb] |

| Quantization | 🟢 Functional | W8A8 available; working on more quantization method support (W4A8, etc) |
| Graph Mode | 🔵 Experimental| Experimental, see detailed note: [vllm-ascend#767][graph_mode] |
| Sleep Mode | 🟢 Functional | |
| Quantization | 🟢 Functional | See detailed note: [Quantization Guide][qaunt] |
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

high

There's a typo in the link reference for the "Quantization Guide". It's written as [qaunt], but the link is defined as [quant]. This results in a broken link.

Suggested change
| Quantization | 🟢 Functional | See detailed note: [Quantization Guide][qaunt] |
| Quantization | 🟢 Functional | See detailed note: [Quantization Guide][quant] |

@wangxiyuan wangxiyuan merged commit faf3406 into vllm-project:v0.11.0-dev Dec 17, 2025
7 checks passed
@wangxiyuan wangxiyuan changed the title 5098 [Doc] Fix user guide error for v0.11.0-dev Dec 17, 2025
@wangxiyuan wangxiyuan deleted the 5098 branch January 12, 2026 03:50
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants