-
Notifications
You must be signed in to change notification settings - Fork 73
Pull requests: vllm-project/vllm-gaudi
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Doc changes from main to 0.11.2
documentation
Improvements or additions to documentation
skip-gaudi-tests
#655
opened Nov 27, 2025 by
mhelf-intel
Loading…
Fix for links in docker md
documentation
Improvements or additions to documentation
skip-gaudi-tests
#653
opened Nov 27, 2025 by
PatrykWo
Loading…
[FIX_FOR_VLLM_LATEST] Fix the Attention imports
#652
opened Nov 27, 2025 by
pawel-olejniczak
Loading…
[GAUDISW-228042] Add support for dynamic quant with V scales on hidden dim
#650
opened Nov 27, 2025 by
dudilester
Loading…
Resolve issue with async scheduling when decode and prompt tokens are mixed
#642
opened Nov 26, 2025 by
tianmu-li
Loading…
[CI] GSM8K: don't override lm_eval batch_size & don't disable prefix caching by default
#640
opened Nov 26, 2025 by
kzawora-intel
Loading…
Removing external links from the main page
documentation
Improvements or additions to documentation
skip-gaudi-tests
#638
opened Nov 26, 2025 by
PatrykWo
Loading…
Enable dequant fp8 weights quantized per-channel with compressed-tensor method
#621
opened Nov 24, 2025 by
mandy-li
Loading…
Fix transformers version mismatch causing editable install failure
#618
opened Nov 21, 2025 by
Saiteja-Garlapati
Loading…
Allow building vllm-plugin docker for ubuntu with upstream torch
#613
opened Nov 21, 2025 by
mmuszynskihabana
Loading…
Previous Next
ProTip!
Type g i on any issue or pull request to go back to the issue listing page.