forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 78
Pull requests: HabanaAI/vllm-fork
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
Automatic Prefix Caching - ux
habana
Issues or PRs submitted by Habana Labs
#902
opened Mar 10, 2025 by
adobrzyn
Loading…
Synchronize vLLM flags to support cross-node inference
#897
opened Mar 7, 2025 by
IT-Forrest
Loading…
[SW-221458] Synchronization between HPU and CPU for more precise TTFT measurement
#892
opened Mar 6, 2025 by
yuwenzho
Loading…
Bump jinja2 from 3.1.4 to 3.1.6
dependencies
Pull requests that update a dependency file
python
Pull requests that update python code
#891
opened Mar 6, 2025 by
dependabot
bot
Loading…
Added the logic to fix the warmup phase for spec decoding when enforce_eager is not used
#880
opened Feb 28, 2025 by
pallavijaini0525
Loading…
[Gaudi][Model] Qwen2.5-vl
New Model
Issue o PR to enable a new model
#870
opened Feb 26, 2025 by
malkomes
Loading…
Update requirements-hpu.txt for open telemetry tracing support
#857
opened Feb 21, 2025 by
louie-tsai
Loading…
enable multi-modal embedding for TIGER-Lab/VLM2Vec-Full T+I on HPU
#854
opened Feb 20, 2025 by
libinta
Loading…
Draft: Another attempt at v1 HPU integration
#831
opened Feb 14, 2025 by
kzawora-intel
•
Draft
22 of 24 tasks
Extend accuracy tests for models that we support
#824
opened Feb 13, 2025 by
AnetaKaczynska
Loading…
Update documentation to reflect current bucket defaults
#817
opened Feb 12, 2025 by
nngokhale
Loading…
Support qwenvl model for HPU
New Model
Issue o PR to enable a new model
#793
opened Feb 7, 2025 by
yingjie-han
Loading…
[DEEPSEEK_V3/R1] includes features of fp8 dequant, MLA, Expert parallelism
#792
opened Feb 6, 2025 by
xuechendi
Loading…
[DO NOT MERGE][PoC] Mark dynamic shapes in torch.compile mode
#755
opened Jan 29, 2025 by
kzawora-intel
•
Draft
Previous Next
ProTip!
no:milestone will show everything without a milestone.