Skip to content

Commit e1abc2d

Browse files
committed
Merge remote-tracking branch 'origin/main'
2 parents db4de44 + a55c728 commit e1abc2d

File tree

9 files changed

+59
-970
lines changed

9 files changed

+59
-970
lines changed

.github/workflows/vllm_ascend_test_pr_light.yaml

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -135,10 +135,12 @@ jobs:
135135
pytest -sv --cov --cov-report=xml:unittests-coverage.xml tests/ut \
136136
--ignore tests/ut/torchair/models/test_torchair_deepseek_mtp.py \
137137
--ignore tests/ut/torchair/models/test_torchair_deepseek_v2.py \
138-
--ignore tests/ut/models/test_qwen2_vl.py \
139-
--ignore tests/ut/models/test_qwen2_5_vl.py \
140-
--ignore tests/ut/models/test_qwen2_5_vl_without_padding.py \
141-
--ignore tests/ut/model_loder
138+
--ignore tests/ut/model_loader/netloader/test_netloader_elastic.py \
139+
--ignore tests/ut/kv_connector/test_remote_prefill_lifecycle.py \
140+
--ignore tests/ut/kv_connector/test_remote_decode_lifecycle.py \
141+
--ignore tests/ut/kv_connector/test_llmdatadist_connector.py \
142+
--ignore tests/ut/ops/test_linear.py \
143+
--ignore tests/ut/core/test_scheduler_dynamic_batch.py
142144
143145
- name: Upload coverage to Codecov
144146
# only upload coverage when commits merged

tests/e2e/nightly/ops/test_fused_moe.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,8 @@
2828
import torch_npu
2929
from vllm.model_executor.layers.activation import SiluAndMul
3030

31-
from vllm_ascend.ops.fused_moe.experts_selector import select_experts
31+
from vllm_ascend.ops.fused_moe.experts_selector import (
32+
check_npu_moe_gating_top_k, select_experts)
3233
from vllm_ascend.ops.fused_moe.moe_mlp import unified_apply_mlp
3334
from vllm_ascend.ops.fused_moe.token_dispatcher import \
3435
TokenDispatcherWithAllGather
@@ -303,7 +304,10 @@ def test_select_experts(
303304
e_score_correction_bias=e_score_correction_bias,
304305
)
305306

306-
if use_grouped_topk:
307+
call_moe_gatingtopk = check_npu_moe_gating_top_k(
308+
hidden_states, topk, topk_group, num_expert_group, scoring_func,
309+
custom_routing_function)
310+
if not call_moe_gatingtopk and use_grouped_topk:
307311
mock_native_grouped_topk.assert_called_once()
308312
else:
309313
mock_native_grouped_topk.assert_not_called()

tests/ut/core/test_schedule_config.py

Lines changed: 0 additions & 134 deletions
This file was deleted.

0 commit comments

Comments
 (0)