Skip to content

Pull requests: vllm-project/vllm

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Sort

Pull requests list

[fix] Correct num_accepted_tokens counting
#10604 opened Nov 24, 2024 by KexinFeng Loading…
[Interleaved ATTN] Support for Mistral-8B
#10591 opened Nov 23, 2024 by patrickvonplaten Loading…
fix json serialization issue frontend
#10580 opened Nov 22, 2024 by maxdebayser Loading…
[Docs] Add dedicated tool calling page to docs documentation Improvements or additions to documentation
#10554 opened Nov 21, 2024 by mgoin Loading…
Add Sageattention backend
#10532 opened Nov 21, 2024 by flozi00 Loading…
[Model]: Add support for Aria model documentation Improvements or additions to documentation
#10514 opened Nov 21, 2024 by xffxff Loading…
[Model] Add OLMo November 2024 model documentation Improvements or additions to documentation
#10503 opened Nov 20, 2024 by 2015aroras Loading…
[Core] Implement disagg prefill by StatelessProcessGroup ci/build ready ONLY add when PR is ready to merge/full CI is needed
#10502 opened Nov 20, 2024 by KuntaiDu Loading…
Support softcap in ROCm Flash Attention
#10500 opened Nov 20, 2024 by hliuca Loading…
ProTip! What’s not been updated in a month: updated:<2024-10-24.