Skip to content

Pull requests: vllm-project/vllm-gaudi

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Doc changes from main to 0.11.2 documentation Improvements or additions to documentation skip-gaudi-tests
#655 opened Nov 27, 2025 by mhelf-intel Loading…
Enabling qwen3-0.6b server/benchmark on a docker
#654 opened Nov 27, 2025 by PatrykWo Loading…
Fix for links in docker md documentation Improvements or additions to documentation skip-gaudi-tests
#653 opened Nov 27, 2025 by PatrykWo Loading…
Trigger CI test to verigy VLLM NIXL functionality
#649 opened Nov 27, 2025 by amathewc Loading…
platform: optimize grouped topk op
#647 opened Nov 27, 2025 by xinyu-intel Loading…
make mla weight contiguous
#646 opened Nov 27, 2025 by xinyu-intel Loading…
bucket: add query len 1 to prefill bucket
#645 opened Nov 27, 2025 by xinyu-intel Loading…
Hybrid KV cache for hpu
#644 opened Nov 26, 2025 by michalkuligowski Draft
Removing external links from the main page documentation Improvements or additions to documentation skip-gaudi-tests
#638 opened Nov 26, 2025 by PatrykWo Loading…
Fix filter for edge case & prefill bs > 1
#634 opened Nov 26, 2025 by adobrzyn Loading…
Fix LoRA tests
#630 opened Nov 25, 2025 by vivekgoe Loading…
Spec decode warmup support
#624 opened Nov 25, 2025 by jerrychenhf Loading…
Fix environment setup for FP8
#623 opened Nov 25, 2025 by yiliu30 Loading…
enable spec decode for Unified Attention, part1
#619 opened Nov 21, 2025 by xuechendi Loading…
lora, fix for PR28545
#617 opened Nov 21, 2025 by iboiko-habana Loading…
Sleep mode support
#584 opened Nov 18, 2025 by Kacper-Pietkun Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.