Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Label
Filter by label
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Milestones
Filter by milestone
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

cache image build
#5419 opened Jun 11, 2024 by khluu Loading…
[Bugfix]Fix evict v2 with long context length
#5411 opened Jun 11, 2024 by puf147 Loading…
[Speculative decoding] Initial spec decode docs
#5400 opened Jun 11, 2024 by cadedaniel Loading…
[WIP][Core][Distributed] add shm broadcast
#5399 opened Jun 11, 2024 by youkaichao Loading…
[Kernel] Vectorized FP8 quantize kernel
#5396 opened Jun 10, 2024 by comaniac Loading…
Set AMD tests soft_fail=false
#5393 opened Jun 10, 2024 by simon-mo Loading…
[CI] Upgrade codespell version.
#5381 opened Jun 10, 2024 by rkooo567 Loading…
[Hardware][Intel] OpenVINO vLLM backend
#5379 opened Jun 10, 2024 by ilya-lavrenov Loading…
[Core][Distributed] add same-node detection
#5369 opened Jun 10, 2024 by youkaichao Loading…
[Core][Bugfix]: fix prefix caching for blockv2
#5364 opened Jun 9, 2024 by leiwen83 Loading…
[Model][Bugfix] Add GLM-4v support
#5358 opened Jun 8, 2024 by songxxzp Loading…
ProTip! Type g i on any issue or pull request to go back to the issue listing page.