diff --git a/vllm/model_executor/layers/fused_moe/fused_moe.py b/vllm/model_executor/layers/fused_moe/fused_moe.py index b750fc713b43f..ecab77a8b6dfb 100644 --- a/vllm/model_executor/layers/fused_moe/fused_moe.py +++ b/vllm/model_executor/layers/fused_moe/fused_moe.py @@ -423,6 +423,11 @@ def fused_experts(hidden_states: torch.Tensor, M, _ = hidden_states.shape E, N, _ = w1.shape + if M > 65536: + # https://github.com/vllm-project/vllm/issues/5938 + raise ValueError("MoE kernel does not support more than 65536 tokens, " + f"but got {M}") + if override_config: config = override_config else: