Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[Bugfix] Fix VLLM_USE_MODELSCOPE issue
#13384 opened Feb 17, 2025 by r4ntix Loading…
[VLM] Check required fields before initializing field config in DictEmbeddingItems documentation Improvements or additions to documentation ready ONLY add when PR is ready to merge/full CI is needed
#13380 opened Feb 17, 2025 by DarkLight1337 Loading…
[MISC] tiny fixes ready ONLY add when PR is ready to merge/full CI is needed
#13378 opened Feb 17, 2025 by MengqingCao Loading…
[V1] Support bad_words in sampler v1
#13376 opened Feb 17, 2025 by 22quinn Draft
1 task
set chunked_prefill off when use mla
#13374 opened Feb 17, 2025 by DragonFive Loading…
[Bugfix] fix xpu communicator
#13368 opened Feb 17, 2025 by yma11 Loading…
[Quant] Arctic SupportsQuant
#13366 opened Feb 17, 2025 by kylesayrs Loading…
Make log statistics interval configurable v1
#13356 opened Feb 16, 2025 by Sakalya Loading…
[Benchmark] Add LongBench to benchmark_serving
#13350 opened Feb 16, 2025 by YuhanLiu11 Loading…
[V1] Get input tokens from scheduler ready ONLY add when PR is ready to merge/full CI is needed v1
#13339 opened Feb 15, 2025 by WoosukKwon Loading…
[Quant] Molmo SupportsQuant
#13336 opened Feb 15, 2025 by kylesayrs Loading…
[Core] Faster logit_bias_logits_processor frontend
#13334 opened Feb 15, 2025 by xu-song Loading…
[Kernel] moe wna16 cuda kernel ci/build
#13321 opened Feb 15, 2025 by jinzhen-lin Loading…
[Model] Add support for GraniteMoeShared models
#13313 opened Feb 15, 2025 by tjohnson31415 Loading…
ProTip! Follow long discussions with comments:>50.