Skip to content

Conversation

tianmu-li
Copy link
Contributor

@tianmu-li tianmu-li commented Oct 8, 2025

When dealing with chunked prompt (input sequence length > max_num_batched_tokens), sometimes there is no output token due to the chunked prompt, but the scheduler expects one. This PR addresses this issue and aligns the behavior with gpu_model_runner.

Copy link

github-actions bot commented Oct 8, 2025

🚧 CI Blocked

The main CI workflow was not started for the following reason:

This is a Draft PR. Please mark it as 'Ready for Review' to trigger the CI.

@tianmu-li tianmu-li force-pushed the async_scheduling_chunk_fix branch from 1f87a51 to 09c183d Compare October 8, 2025 18:32
Copy link

github-actions bot commented Oct 8, 2025

🚧 CI Blocked

The main CI workflow was not started for the following reason:

This is a Draft PR. Please mark it as 'Ready for Review' to trigger the CI.

@tianmu-li tianmu-li marked this pull request as ready for review October 8, 2025 18:42
@tianmu-li tianmu-li changed the title [WIP] Fix issue with async_scheduling when dealing with chunked input Fix issue with async_scheduling when dealing with chunked input Oct 8, 2025
Signed-off-by: Tianmu Li <[email protected]>
@tianmu-li
Copy link
Contributor Author

@afierka-intel Can you check if this fixes the issue with long context? It works from my tests.

@afierka-intel
Copy link
Collaborator

/run-gaudi-tests

1 similar comment
@michalkuligowski
Copy link
Collaborator

/run-gaudi-tests

Copy link
Collaborator

@michalkuligowski michalkuligowski left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please cherrypick to 0.11 also

@michalkuligowski
Copy link
Collaborator

/run-gaudi-tests

Copy link

github-actions bot commented Oct 9, 2025

✅ CI Passed

All checks passed successfully against the following vllm commit:
da3fa78dc98f3001e5fb703729a77311146e0cd3

@michalkuligowski
Copy link
Collaborator

/run-gaudi-tests

@michalkuligowski
Copy link
Collaborator

/run-gaudi-tests

Copy link

✅ CI Passed

All checks passed successfully against the following vllm commit:
da3fa78dc98f3001e5fb703729a77311146e0cd3

@michalkuligowski
Copy link
Collaborator

/run-gaudi-tests

Copy link

✅ CI Passed

All checks passed successfully against the following vllm commit:
da3fa78dc98f3001e5fb703729a77311146e0cd3

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants