CVE-2025-49847

llama.cpp is an inference of several LLM models in C/C++. Prior to version b5662, an attacker-supplied GGUF model vocabulary can trigger a buffer overflow in llama.cpp’s vocabulary-loading code. Specifically, the helper _try_copy in llama.cpp/src/vocab.cpp: llama_vocab::impl::token_to_piece() casts a very large size_t token length into an int32_t, causing the length check (if (length < (int32_t)size)) to be bypassed. As a result, memcpy is still called with that oversized size, letting a malicious model overwrite memory beyond the intended buffer. This can lead to arbitrary memory corruption and potential code execution. This issue has been patched in version b5662.
CVSS

No CVSS.

Configurations

No configuration.

History

17 Jun 2025, 20:15

Type Values Removed Values Added
New CVE

Information

Published : 2025-06-17 20:15

Updated : 2025-06-17 20:50


NVD link : CVE-2025-49847

Mitre link : CVE-2025-49847


JSON object : View

Products Affected

No product.

CWE
CWE-195

Signed to Unsigned Conversion Error

CWE-119

Improper Restriction of Operations within the Bounds of a Memory Buffer