-
-
Notifications
You must be signed in to change notification settings - Fork 12.4k
Pull requests: vllm-project/vllm
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
[Misc] Add VLLM_USE_FLASHINFER_ROPE to control the RoPE kernel for cuda
nvidia
#31893
opened Jan 7, 2026 by
elvischenv
Loading…
5 tasks
[BugFix][Model]: Pass indexer_rotary_emb to MLA modules
#31892
opened Jan 7, 2026 by
aaarkai
Loading…
3 of 5 tasks
[Chore] Migrate V0 attention utils
ready
ONLY add when PR is ready to merge/full CI is needed
rocm
Related to AMD ROCm
v1
#31891
opened Jan 7, 2026 by
DarkLight1337
Loading…
5 tasks
[Models] Allow converting Qwen3-VL into Reranker model
documentation
Improvements or additions to documentation
qwen
Related to Qwen models
[Misc] Unify
get_kv_cache_stride_order code style
nvidia
v1
#31887
opened Jan 7, 2026 by
NickLucche
Loading…
[Feature][Bugfix] Support draft model tp any of speculative decode
speculative-decoding
v1
#31886
opened Jan 7, 2026 by
stormchasingg
Loading…
[Frontend] reimplement beam-search reference transformers
frontend
#31882
opened Jan 7, 2026 by
satoaoki1999
Loading…
3 of 5 tasks
[Feature][Benchmarks] Custom dataset: read output length from dataset
performance
Performance-related issues
#31881
opened Jan 7, 2026 by
sducouedic
Loading…
[Misc] Set default torch num threads for input processing
ready
ONLY add when PR is ready to merge/full CI is needed
v1
#31879
opened Jan 7, 2026 by
ywang96
Loading…
5 tasks
[Feature] Add flag to disable FlashInfer autotune
#31875
opened Jan 7, 2026 by
mmangkad
Loading…
5 tasks
[Feature] Add command-line argument support to basic.py example
documentation
Improvements or additions to documentation
#31874
opened Jan 7, 2026 by
liangzhang-keepmoving
Loading…
3 of 5 tasks
[CI][BugFix][AMD] Actually skip tests marked @pytest.mark.skip_v1
ci/build
rocm
Related to AMD ROCm
#31873
opened Jan 7, 2026 by
rasmith
Loading…
[OpenAI] Extend VLLMValidationError to additional validation parameters
frontend
ready
ONLY add when PR is ready to merge/full CI is needed
v1
#31870
opened Jan 7, 2026 by
R3hankhan123
Loading…
5 tasks
[Doc] Fix: Correct vLLM announcing blog post link in docs
documentation
Improvements or additions to documentation
#31868
opened Jan 7, 2026 by
Ayobami-00
Loading…
5 tasks
[Bugfix] Supplement CpuCommunicator with dispatch and combine im…
cpu
Related to CPU backends
#31867
opened Jan 7, 2026 by
kzwrime
Loading…
3 of 5 tasks
[refactor] refactor memory constants usage
cpu
Related to CPU backends
multi-modality
Related to multi-modality (#4194)
v1
#31865
opened Jan 7, 2026 by
andyxning
Loading…
5 tasks
feat: add support for logging to file via VLLM_LOG_FILE env var
#31863
opened Jan 7, 2026 by
leejianwoo-collab
Loading…
5 tasks
[Bugfix] manully free encode cache of waiting request to avoid potential dead …
v1
#31857
opened Jan 7, 2026 by
frelam
Loading…
5 tasks
Add GPU memory usage warning system
documentation
Improvements or additions to documentation
v1
#31853
opened Jan 7, 2026 by
Dedulus
Loading…
5 tasks done
[Attention] Full CG support for llama4 and remove use of deprecated properties
llama
Related to Llama models
v1
#31852
opened Jan 7, 2026 by
LucasWilkinson
•
Draft
5 tasks
[CI] Fix weight mapping test for transformers v5 tied weights
multi-modality
Related to multi-modality (#4194)
#31849
opened Jan 7, 2026 by
AndreasKaratzas
Loading…
[Model] Add Grok-2
documentation
Improvements or additions to documentation
new-model
Requests to new models
#31847
opened Jan 7, 2026 by
dangoldbj
Loading…
5 tasks
[4/n] Migrate pos_encoding sampler and fused_qknorm_rope to libtorch stable ABI
ci/build
cpu
Related to CPU backends
nvidia
#31842
opened Jan 6, 2026 by
mikaylagawarecki
•
Draft
5 tasks
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.