Skip to content

Commit e850c8b

Browse files
committed
fix merge
Signed-off-by: Bill Nell <[email protected]>
1 parent 4863dff commit e850c8b

File tree

1 file changed

+2
-5
lines changed
  • vllm/model_executor/layers/quantization

1 file changed

+2
-5
lines changed

vllm/model_executor/layers/quantization/fp8.py

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1171,12 +1171,9 @@ def apply(
11711171
x: torch.Tensor,
11721172
router_logits: torch.Tensor,
11731173
) -> torch.Tensor | tuple[torch.Tensor, torch.Tensor]:
1174-
if layer.enable_eplb:
1175-
assert layer.expert_load_view is not None
1176-
assert layer.logical_to_physical_map is not None
1177-
assert layer.logical_replica_count is not None
1178-
11791174
if self.flashinfer_moe_backend == FlashinferMoeBackend.TENSORRT_LLM:
1175+
if layer.enable_eplb:
1176+
raise NotImplementedError("EPLB not supported for `Fp8MoEMethod` yet.")
11801177
assert layer.activation == "silu", (
11811178
f"Expected 'silu' activation but got {layer.activation}"
11821179
)

0 commit comments

Comments
 (0)