Skip to content

Commit c6d51d2

Browse files
committed
Fix condition for expert fusion by updating the check for 'enable_deepep_moe' to include 'enable_ep_moe' in DeepseekV2ForCausalLM. sgl-project#6767
1 parent b4ae984 commit c6d51d2

File tree

1 file changed

+6
-1
lines changed

1 file changed

+6
-1
lines changed

python/sglang/srt/models/deepseek_v2.py

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1688,7 +1688,12 @@ def determine_n_share_experts_fusion(
16881688
and torch.cuda.get_device_capability("cuda") >= (9, 0)
16891689
and self.config.architectures[0] == architecture
16901690
and self.config.n_routed_experts == 256
1691-
and (not global_server_args_dict["enable_deepep_moe"])
1691+
and (
1692+
not (
1693+
global_server_args_dict["enable_deepep_moe"]
1694+
or global_server_args_dict["enable_ep_moe"]
1695+
)
1696+
)
16921697
):
16931698
self.n_share_experts_fusion = self.tp_size
16941699
global_server_args_dict["n_share_experts_fusion"] = self.tp_size

0 commit comments

Comments
 (0)