Description
vLLM is a library for LLM inference and serving. vllm/model_executor/weight_utils.py implements hf_model_weights_iterator to load the model checkpoint, which is downloaded from huggingface. It uses the torch.load function and the weights_only parameter defaults to False. When torch.load loads malicious pickle data, it will execute arbitrary code during unpickling. This vulnerability is fixed in v0.7.0.
References (4)
Core 4
Core References
Vendor Advisory x_refsource_confirm
https://github.com/vllm-project/vllm/security/advisories/GHSA-rh4j-5rhw-hr54
Issue Tracking, Patch x_refsource_misc
https://github.com/vllm-project/vllm/pull/12366
Patch x_refsource_misc
https://github.com/vllm-project/vllm/commit/d3d6bb13fb62da3234addf6574922a4ec0513d04
Technical Description x_refsource_misc
https://pytorch.org/docs/stable/generated/torch.load.html
Scores
CVSS v3
7.5
EPSS
0.0101
EPSS Percentile
77.2%
Attack Vector
NETWORK
CVSS:3.1/AV:N/AC:H/PR:N/UI:R/S:U/C:H/I:H/A:H
CISA SSVC
Vulnrichment
Exploitation
none
Automatable
no
Technical Impact
total
Details
CWE
CWE-502
Status
published
Products (2)
pypi/vllm
0 - 0.7.0PyPI
vllm/vllm
< 0.7.0
Published
Jan 27, 2025
Tracked Since
Feb 18, 2026