CVE-2025-62426

MEDIUM

Vllm < 0.11.1 - Resource Allocation Without Limits

Title source: rule
STIX 2.1

Description

vLLM is an inference and serving engine for large language models (LLMs). From version 0.5.5 to before 0.11.1, the /v1/chat/completions and /tokenize endpoints allow a chat_template_kwargs request parameter that is used in the code before it is properly validated against the chat template. With the right chat_template_kwargs parameters, it is possible to block processing of the API server for long periods of time, delaying all other requests. This issue has been patched in version 0.11.1.

Scores

CVSS v3 6.5
EPSS 0.0009
EPSS Percentile 25.1%
Attack Vector NETWORK
CVSS:3.1/AV:N/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H

CISA SSVC

Vulnrichment
Exploitation none
Automatable no
Technical Impact partial

Details

CWE
CWE-770
Status published
Products (3)
pypi/vllm 0.5.5 - 0.11.1PyPI
vllm/vllm 0.11.1 rc0 (2 CPE variants)
vllm/vllm 0.5.5 - 0.11.1
Published Nov 21, 2025
Tracked Since Feb 18, 2026