Skip to content

Commit 1bcc5bc

Browse files
authored
feat: Update llama.cpp to ggml-org/llama.cpp@3bd9aa1f9 (#2176)
* feat: Update llama.cpp to ggml-org/llama.cpp@3bd9aa1f9 * docs: Update changelog for llama.cpp bump
1 parent 02d6bee commit 1bcc5bc

File tree

3 files changed

+9
-3
lines changed

3 files changed

+9
-3
lines changed

CHANGELOG.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,8 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
77

88
## [Unreleased]
99

10+
- feat: Update llama.cpp to ggerganov/llama.cpp@3bd9aa1f9 and sync Python bindings
11+
1012
## [0.3.20]
1113

1214
- refactor: Replace deprecated llama.cpp references in library, docs, and examples by @abetlen in #2170

llama_cpp/llama_cpp.py

Lines changed: 6 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,8 @@ def _warn_deprecated(symbol: str, hint: str) -> None:
8888
# GGML_TYPE_IQ1_M = 29,
8989
# GGML_TYPE_MXFP4 = 39,
9090
# GGML_TYPE_NVFP4 = 40,
91-
# GGML_TYPE_COUNT,
91+
# GGML_TYPE_Q1_0 = 41,
92+
# GGML_TYPE_COUNT = 42,
9293
# };
9394
GGML_TYPE_F32 = 0
9495
GGML_TYPE_F16 = 1
@@ -120,7 +121,8 @@ def _warn_deprecated(symbol: str, hint: str) -> None:
120121
GGML_TYPE_IQ1_M = 29
121122
GGML_TYPE_MXFP4 = 39
122123
GGML_TYPE_NVFP4 = 40
123-
GGML_TYPE_COUNT = 41
124+
GGML_TYPE_Q1_0 = 41
125+
GGML_TYPE_COUNT = 42
124126

125127
# from ggml-backend.h
126128
# typedef bool (*ggml_backend_sched_eval_callback)(struct ggml_tensor * t, bool ask, void * user_data);
@@ -406,6 +408,7 @@ def _warn_deprecated(symbol: str, hint: str) -> None:
406408
# LLAMA_FTYPE_MOSTLY_TQ2_0 = 37, // except 1d tensors
407409
# LLAMA_FTYPE_MOSTLY_MXFP4_MOE = 38, // except 1d tensors
408410
# LLAMA_FTYPE_MOSTLY_NVFP4 = 39, // except 1d tensors
411+
# LLAMA_FTYPE_MOSTLY_Q1_0 = 40, // except 1d tensors
409412
#
410413
# LLAMA_FTYPE_GUESSED = 1024, // not specified in the model file
411414
# };
@@ -446,6 +449,7 @@ def _warn_deprecated(symbol: str, hint: str) -> None:
446449
LLAMA_FTYPE_MOSTLY_TQ2_0 = 37
447450
LLAMA_FTYPE_MOSTLY_MXFP4_MOE = 38
448451
LLAMA_FTYPE_MOSTLY_NVFP4 = 39
452+
LLAMA_FTYPE_MOSTLY_Q1_0 = 40
449453
LLAMA_FTYPE_GUESSED = 1024
450454

451455
# enum llama_rope_scaling_type {

vendor/llama.cpp

0 commit comments

Comments
 (0)