-
-
Notifications
You must be signed in to change notification settings - Fork 11.2k
[Model] Apply shared experts overlap optimization to all models with shared experts #26145
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Signed-off-by: Bill Nell <[email protected]>
Signed-off-by: Bill Nell <[email protected]>
Signed-off-by: Bill Nell <[email protected]>
Signed-off-by: Bill Nell <[email protected]>
hmellor
reviewed
Oct 7, 2025
Signed-off-by: Bill Nell <[email protected]>
mgoin
approved these changes
Oct 8, 2025
Member
mgoin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM, nice refactor to get thank you
Signed-off-by: Bill Nell <[email protected]>
Signed-off-by: Bill Nell <[email protected]>
yang926
pushed a commit
to yang926/vllm_1008
that referenced
this pull request
Oct 9, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: yang926 <[email protected]>
yang926
pushed a commit
to yang926/vllm_1008
that referenced
this pull request
Oct 9, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: yang926 <[email protected]>
xuebwang-amd
pushed a commit
to xuebwang-amd/vllm
that referenced
this pull request
Oct 10, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: xuebwang-amd <[email protected]>
Dhruvilbhatt
pushed a commit
to Dhruvilbhatt/vllm
that referenced
this pull request
Oct 14, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: Dhruvil Bhatt <[email protected]>
This was referenced Oct 16, 2025
lywa1998
pushed a commit
to lywa1998/vllm
that referenced
this pull request
Oct 20, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]>
alhridoy
pushed a commit
to alhridoy/vllm
that referenced
this pull request
Oct 24, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]>
wangxiyuan
pushed a commit
to vllm-project/vllm-ascend
that referenced
this pull request
Oct 24, 2025
### What this PR does / why we need it? This is the step 1 of refactoring code to adapt with vllm main, and this pr aligned with vllm-project/vllm@17c540a 1. refactor deepseek to the latest code arch as of vllm-project/vllm@17c540a 2. bunches of fixes due to vllm changes - Fix `AscendScheduler` `__post_init__`, caused by vllm-project/vllm#25075 - Fix `AscendScheduler` init got an unexpected arg `block_size`, caused by vllm-project/vllm#26296 - Fix `KVCacheManager` `get_num_common_prefix_blocks` arg, caused by vllm-project/vllm#23485 - Fix `MLAAttention` import,caused by vllm-project/vllm#25103 - Fix `SharedFusedMoE` import, caused by vllm-project/vllm#26145 - Fix `LazyLoader` improt, caused by vllm-project/vllm#27022 - Fix `vllm.utils.swap_dict_values` improt, caused by vllm-project/vllm#26990 - Fix `Backend` enum import, caused by vllm-project/vllm#25893 - Fix `CompilationLevel` renaming to `CompilationMode` issue introduced by vllm-project/vllm#26355 - Fix fused_moe ops, caused by vllm-project/vllm#24097 - Fix bert model because of `inputs_embeds`, caused by vllm-project/vllm#25922 - Fix MRope because of `get_input_positions_tensor` to `get_mrope_input_positions`, caused by vllm-project/vllm#24172 - Fix `splitting_ops` changes introduced by vllm-project/vllm#25845 - Fix multi-modality changes introduced by vllm-project/vllm#16229 - Fix lora bias dropping issue introduced by vllm-project/vllm#25807 - Fix structured ouput break introduced by vllm-project/vllm#26737 ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? CI passed with existing test. - vLLM version: v0.11.0rc3 - vLLM main: https://github.com/vllm-project/vllm/commit/v0.11.0 --------- Signed-off-by: MengqingCao <[email protected]> Signed-off-by: Icey <[email protected]> Co-authored-by: Icey <[email protected]>
xuebwang-amd
pushed a commit
to xuebwang-amd/vllm
that referenced
this pull request
Oct 24, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: xuebwang-amd <[email protected]>
0xrushi
pushed a commit
to 0xrushi/vllm
that referenced
this pull request
Oct 26, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: 0xrushi <[email protected]>
0xrushi
pushed a commit
to 0xrushi/vllm
that referenced
this pull request
Oct 26, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]> Signed-off-by: 0xrushi <[email protected]>
rtourgeman
pushed a commit
to rtourgeman/vllm
that referenced
this pull request
Nov 10, 2025
…shared experts (vllm-project#26145) Signed-off-by: Bill Nell <[email protected]>
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Purpose
SharedFusedMoEin all models that use shared experts. This will enable the shared experts/communication overlap optimization for all the changed models.SharedFusedMoEclass tofused_moedirectory.SharedFusedMoEto behave likeFusedMoEwhen theshared_expertsareNonetorch.compile.For most models the changes consist of renaming
FusedMoE->SharedFusedMoEand passing the shared experts module as a parameter toSharedFusedMoE. A few models required extra tweaks:aria,ernie45_vl_moeand theqwenmodels.Test Plan
Test all modified models.
Note: all model types appear to be covered by tests/models/registry.py
Test Result
TBD