Skip to content

Pull requests: vllm-project/vllm-gaudi

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

[SW-243111] Add correctors for decode buckets
#509 opened Oct 31, 2025 by jbyczkow Loading…
DP: allreduce on the host
#498 opened Oct 29, 2025 by xinyu-intel Loading…
Udpate TESTOWNERS
#495 opened Oct 28, 2025 by jbyczkow Loading…
Documentation updates - part 1 documentation Improvements or additions to documentation skip-gaudi-tests
#493 opened Oct 28, 2025 by mhelf-intel Loading…
Initial Commit GPT-OSS
#485 opened Oct 28, 2025 by hlahkar Loading…
[SW-242523] Support per-tensor FP8 scaling
#483 opened Oct 27, 2025 by skavulya Loading…
Fix docker cmdlines for v0.10.2_next workarounds
#477 opened Oct 25, 2025 by nngokhale Loading…
Simplify requirements
#458 opened Oct 23, 2025 by pawel-olejniczak Loading…
Enable triangular mask with valid_seq_lengths
#454 opened Oct 23, 2025 by kamil-kaczor Loading…
Remove VLLM_DELAYED_SAMPLING
#433 opened Oct 21, 2025 by xwu-intel Loading…
enable gdr on 10.2 baseline
#431 opened Oct 20, 2025 by hsubramony Draft
Fix for Llama4 static quantization
#430 opened Oct 20, 2025 by vidyasiv Loading…
multimodal support for unified attn
#423 opened Oct 19, 2025 by attafosu Loading…
Fix docker cmdlines for v.0.11.0 work arounds
#417 opened Oct 17, 2025 by nngokhale Loading…
Add async scheduling for unified attention
#414 opened Oct 16, 2025 by tianmu-li Loading…
ProTip! Find all pull requests that aren't related to any open issues with -linked:issue.