-
Notifications
You must be signed in to change notification settings - Fork 0
fix(deps): update dependency vllm to v0.11.0 #11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
870b86a to
8177d49
Compare
|
No applications have been configured for previews targeting branch: main. To do so go to restack console and configure your applications for previews. |
8177d49 to
7653ea9
Compare
7653ea9 to
b163b3b
Compare
b163b3b to
fd2265d
Compare
|
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
|
fd2265d to
233f41a
Compare
233f41a to
0ab9c3b
Compare
0ab9c3b to
28f6622
Compare
28f6622 to
bcef6f4
Compare
bcef6f4 to
9c2649d
Compare
9c2649d to
da9cf63
Compare
da9cf63 to
adfc43c
Compare
adfc43c to
f961c1b
Compare
f961c1b to
ca1b26b
Compare
ca1b26b to
ad11c89
Compare
This PR contains the following updates:
==0.7.2->==0.11.0Release Notes
vllm-project/vllm (vllm)
v0.11.0Compare Source
Highlights
This release features 538 commits, 207 contributors (65 new contributors)!
Note: In v0.11.0 (and v0.10.2),
--async-schedulingwill produce gibberish output in some cases such as preemption and others. This functionality is correct in v0.10.1. We are actively fixing it for the next version.Model Support
Engine Core
Hardware & Performance
Large Scale Serving & Performance
Quantization
API & Frontend
Security
Dependencies
xm.mark_stepin favor oftorch_xla.sync(#25254).V0 Deprecation
What's Changed
cpu_attn.py:_run_sdpa_forwardfor better memory access by @ignaciosica in #24701--enable-log-outputsdoes not match the documentation by @kebe7jun in #24626_validate_and_reshape_mm_tensorby @lgeiger in #24742supports_kwby @lgeiger in #24773s3_utilstype hints withBaseClientby @Zerohertz in #24825stopin reasoning content by @gaocegege in #14550kv_output_aggregatorsupport heterogeneous by @LCAIZJ in #23917MultiModalConfigfromconfig/__init__.pytoconfig/multimodal.pyby @hmellor in #24659HuggingFace->Hugging FaceinIntegration with Hugging Facedocs by @sergiopaniego in #24889is_flashmla_supportedCheck Error by @yewentao256 in #24774n_groups % tp_size == 0by @tomeras91 in #24593SpeculativeConfigfromconfig/__init__.pytoconfig/speculative.pyby @hmellor in #24904EngineCoreRequestarguments in tests and fix extra kwargs by @qthequartermasterman in #24987CpuGpuBufferfor block table tensors by @njhill in #24795AutoModelForVision2Seqby @DarkLight1337 in #25065cutlass_mlahang by @alexm-redhat in #24966MultiModalCacheby @lgeiger in #25006sliding_windowfrom text config in Gemma3 MM by @hmellor in #25085Configuration
📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox.
🔕 Ignore: Close this PR and you won't be reminded about this update again.
This PR was generated by Mend Renovate. View the repository job log.