forked from vllm-project/vllm
-
Notifications
You must be signed in to change notification settings - Fork 135
Pull requests: HabanaAI/vllm-fork
Author
Label
Projects
Milestones
Reviews
Assignee
Sort
Pull requests list
PD scripts update to upgrade mooncake and enable fp8_inc by default
#2212
opened Jan 7, 2026 by
Yanli2190
Loading…
3 tasks
Fix the error of using the wrong model name in the document.
#2208
opened Dec 29, 2025 by
wenbinc-Bin
Loading…
Fix AttributeError in fixed_sub_image_list for embedding models
#2206
opened Dec 24, 2025 by
majunpo
Loading…
Add FA3 opt in INC and upgrade mooncake to 0.3.7
#2199
opened Dec 18, 2025 by
Yanli2190
Loading…
3 tasks
Delay prefix cache calculation to find longest common prefix
#2170
opened Dec 8, 2025 by
ikurtchen
Loading…
3 tasks done
Bump actions/stale from 9.1.0 to 10.1.1
dependencies
Pull requests that update a dependency file
github_actions
Pull requests that update GitHub Actions code
#2169
opened Dec 8, 2025 by
dependabot
bot
Loading…
Enable delayed sampling for warmup also to remove graph compilation i…
#2164
opened Dec 4, 2025 by
yeonsily
Loading…
3 tasks
add VLLM_ENGINE_PROFILER_SKIP_STEPS to the engine profiler
#2143
opened Nov 19, 2025 by
yangulei
Loading…
[DeepSeek R1] chunked prefill warmup with chunk size
#2135
opened Nov 14, 2025 by
jerrychenhf
Loading…
Workaround for Assertion error when embedding with bge-m3 in lazy mode
#2093
opened Oct 28, 2025 by
slokesha
Loading…
Previous Next
ProTip!
What’s not been updated in a month: updated:<2025-12-09.