CVE-2026-44223
Severity CVSS v4.0:
Pending analysis
Type:
Unavailable / Other
Publication date:
12/05/2026
Last modified:
15/05/2026
Description
vLLM is an inference and serving engine for large language models (LLMs). From to before 0.20.0, the extract_hidden_states speculative decoding proposer in vLLM returns a tensor with an incorrect shape after the first decode step, causing a RuntimeError that crashes the EngineCore process. The crash is triggered when any request in the batch uses sampling penalty parameters (repetition_penalty, frequency_penalty, or presence_penalty). A single request with a penalty parameter (e.g., "repetition_penalty": 1.1) is sufficient to crash the server. This vulnerability is fixed in 0.20.0.
Impact
Base Score 3.x
6.50
Severity 3.x
MEDIUM
Vulnerable products and versions
| CPE | From | Up to |
|---|---|---|
| cpe:2.3:a:vllm:vllm:*:*:*:*:*:*:*:* | 0.18.0 (including) | 0.20.0 (excluding) |
To consult the complete list of CPE names with products and versions, see this page



