We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 086e3c0 commit b6192baCopy full SHA for b6192ba
vllm/v1/worker/gpu_model_runner.py
@@ -287,7 +287,6 @@ def __init__(
287
scheduler_config.prefill_max_num_batched_tokens,
288
)
289
self.dcp_rank = 0 if self.dcp_world_size <= 1 else get_dcp_group().rank_in_group
290
- self.max_num_tokens = scheduler_config.max_num_batched_tokens
291
self.max_num_reqs = scheduler_config.max_num_seqs
292
293
# Broadcast PP output for external_launcher (torchrun)
0 commit comments