Severity
8.8HIGHNVD
EPSS
0.1%
top 67.69%
CISA KEV
Not in KEV
Exploit
No known exploits
Affected products
Timeline
PublishedNov 21
Latest updateJan 8

Description

vLLM is an inference and serving engine for large language models (LLMs). From versions 0.10.2 to before 0.11.1, a memory corruption vulnerability could lead to a crash (denial-of-service) and potentially remote code execution (RCE), exists in the Completions API endpoint. When processing user-supplied prompt embeddings, the endpoint loads serialized tensors using torch.load() without sufficient validation. Due to a change introduced in PyTorch 2.8.0, sparse tensor integrity checks are disabled

CVSS vector

CVSS:3.1/AV:N/AC:L/PR:L/UI:N/S:U/C:H/I:H/A:HExploitability: 2.8 | Impact: 5.9

Affected Packages3 packages

NVDvllm/vllm0.10.20.11.1+1
PyPIvllm/vllm0.10.20.13.0+1
CVEListV5vllm-project/vllm>= 0.10.2, < 0.11.1

Patches

🔴Vulnerability Details

4
OSV
vLLM introduced enhanced protection for CVE-2025-621642026-01-08
GHSA
vLLM introduced enhanced protection for CVE-2025-621642026-01-08
GHSA
vLLM deserialization vulnerability leading to DoS and potential RCE2025-11-20
OSV
vLLM deserialization vulnerability leading to DoS and potential RCE2025-11-20

📋Vendor Advisories

1
Red Hat
vllm: VLLM deserialization vulnerability leading to DoS and potential RCE2025-11-21

🕵️Threat Intelligence

1
Wiz
GHSA-mcmc-2m55-j8jj Impact, Exploitability, and Mitigation Steps | Wiz