CVE-2025-62372
vLLM vulnerability analysis and mitigation

Overview

vLLM is an inference and serving engine for large language models (LLMs). From version 0.5.5 to before 0.11.1, users can crash the vLLM engine serving multimodal models by passing multimodal embedding inputs with correct ndim but incorrect shape (e.g. hidden dimension is wrong), regardless of whether the model is intended to support such inputs. The vulnerability was discovered and disclosed on November 20, 2025, and has been assigned CVE-2025-62372 (GitHub Advisory).

Technical details

The vulnerability stems from insufficient validation of multimodal embedding inputs. The engine only validates the number of dimensions (ndim) of the tensor but not the full shape in the MultiModalDataParser. This leads to two failure scenarios: 1) For models supporting image embedding inputs, the engine crashes when scattering embeddings to inputsembeds due to mismatched shapes, and 2) For models not supporting image embedding inputs, the engine crashes during input validation inside getinput_embeddings. The vulnerability has been assigned a CVSS v3.1 base score of 6.5 (Moderate) with vector AV:N/AC:L/PR:L/UI:N/S:U/C:N/I:N/A:H (GitHub Advisory).

Impact

The vulnerability allows attackers to cause denial of service by crashing the vLLM engine through malformed multimodal embedding inputs. This affects the availability of the model serving infrastructure (GitHub Advisory).

Mitigation and workarounds

Prior to updating, users can mitigate the vulnerability by: 1) Using API keys to limit access to trusted users only, or 2) Setting --limit-mm-per-prompt to 0 for all non-text modalities to ban multimodal inputs, though this defeats the purpose of using multimodal models. The vulnerability has been patched in version 0.11.1 (GitHub Advisory).

Community reactions

The vulnerability was responsibly disclosed and patched through a coordinated effort between the reporter DarkLight1337 and reviewers ywang96 and Isotr0py. The fix was implemented through PR #27204 which introduced additional validation flags for loading text and image embeddings (GitHub PR).

Additional resources


SourceThis report was generated using AI

Related vLLM vulnerabilities:

CVE ID

Severity

Score

Technologies

Component name

CISA KEV exploit

Has fix

Published date

CVE-2025-62164HIGH8.8
  • vLLMvLLM
  • vllm
NoYesNov 21, 2025
CVE-2025-62372HIGH8.3
  • vLLMvLLM
  • vllm
NoYesNov 21, 2025
CVE-2025-6242HIGH7.1
  • ChainguardChainguard
  • py3-vllm-cuda-12.4
NoYesOct 07, 2025
CVE-2025-62426MEDIUM6.5
  • vLLMvLLM
  • vllm
NoYesNov 21, 2025
CVE-2025-61620N/AN/A
  • ChainguardChainguard
  • py3-vllm-cuda-12.4
NoYesOct 08, 2025

Free Vulnerability Assessment

Benchmark your Cloud Security Posture

Evaluate your cloud security practices across 9 security domains to benchmark your risk level and identify gaps in your defenses.

Request assessment

Get a personalized demo

Ready to see Wiz in action?

"Best User Experience I have ever seen, provides full visibility to cloud workloads."
David EstlickCISO
"Wiz provides a single pane of glass to see what is going on in our cloud environments."
Adam FletcherChief Security Officer
"We know that if Wiz identifies something as critical, it actually is."
Greg PoniatowskiHead of Threat and Vulnerability Management