forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 66
Pull requests: HabanaAI/vllm-fork
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Add interleave sliding window by using fusedsdpa kernel.
#725
opened Jan 22, 2025 by
libinta
Loading…
hotfix - Revert vllm/attention/layer.py changes from 0f8cafe - fix torch.compile recompilations
#709
opened Jan 20, 2025 by
RafLit
Loading…
Bump jinja2 from 3.1.4 to 3.1.5
dependencies
Pull requests that update a dependency file
#679
opened Jan 12, 2025 by
dependabot
bot
Loading…
Fix: selecting correct backend for MultiHeadAttention
habana
Issues or PRs submitted by Habana Labs
#645
opened Dec 18, 2024 by
adobrzyniewicz-habana
Loading…
[WIP] Add HPU support to vLLM v1 - cont.
#609
opened Dec 10, 2024 by
kzawora-intel
Loading…
21 of 23 tasks
Add in Dockerfile.hpu.ubi
external
Issues or PRs submitted by external users
#602
opened Dec 9, 2024 by
Xaenalt
Loading…
Previous Next
ProTip!
Updated in the last three days: updated:>2025-01-23.