Skip to content

Commit 587ee3b

Browse files
committed
fix merge
Signed-off-by: Bill Nell <[email protected]>
1 parent b3d76a5 commit 587ee3b

File tree

1 file changed

+2
-5
lines changed
  • vllm/model_executor/layers/quantization

1 file changed

+2
-5
lines changed

vllm/model_executor/layers/quantization/fp8.py

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1143,12 +1143,9 @@ def apply(
11431143
x: torch.Tensor,
11441144
router_logits: torch.Tensor,
11451145
) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]:
1146-
if layer.enable_eplb:
1147-
assert layer.expert_load_view is not None
1148-
assert layer.logical_to_physical_map is not None
1149-
assert layer.logical_replica_count is not None
1150-
11511146
if self.flashinfer_moe_backend == FlashinferMoeBackend.TENSORRT_LLM:
1147+
if layer.enable_eplb:
1148+
raise NotImplementedError("EPLB not supported for `Fp8MoEMethod` yet.")
11521149
assert layer.activation == "silu", (
11531150
f"Expected 'silu' activation but got {layer.activation}"
11541151
)

0 commit comments

Comments
 (0)