Skip to content

[V1][Bugfix] Standardize quantized kv cache rejection for attention backends#14221

Merged
simon-mo merged 4 commits intovllm-project:mainfrom
neuralmagic:reject-fp8-kv-cache-for-unsupported-attn
Mar 6, 2025
Merged

[V1][Bugfix] Standardize quantized kv cache rejection for attention backends#14221
simon-mo merged 4 commits intovllm-project:mainfrom
neuralmagic:reject-fp8-kv-cache-for-unsupported-attn

Commits

Commits on Mar 4, 2025

Commits on Mar 6, 2025