Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Commit bb610b9

Browse files
committed
Update llama.cpp
1 parent f015966 commit bb610b9

File tree

2 files changed

+6
-1
lines changed

2 files changed

+6
-1
lines changed

llama_cpp/llama_cpp.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -180,6 +180,8 @@ def _load_shared_library(lib_base_name: str):
180180
# LLAMA_FTYPE_MOSTLY_Q5_K_M = 17, // except 1d tensors
181181
# LLAMA_FTYPE_MOSTLY_Q6_K = 18, // except 1d tensors
182182
# LLAMA_FTYPE_MOSTLY_IQ2_XXS = 19, // except 1d tensors
183+
# LLAMA_FTYPE_MOSTLY_IQ2_XS = 20, // except 1d tensors
184+
# LLAMA_FTYPE_MOSTLY_Q2_K_S = 21, // except 1d tensors
183185

184186
# LLAMA_FTYPE_GUESSED = 1024, // not specified in the model file
185187
# };
@@ -200,6 +202,9 @@ def _load_shared_library(lib_base_name: str):
200202
LLAMA_FTYPE_MOSTLY_Q5_K_S = 16
201203
LLAMA_FTYPE_MOSTLY_Q5_K_M = 17
202204
LLAMA_FTYPE_MOSTLY_Q6_K = 18
205+
LLAMA_FTYPE_MOSTLY_IQ2_XXS = 19
206+
LLAMA_FTYPE_MOSTLY_IQ2_XS = 20
207+
LLAMA_FTYPE_MOSTLY_Q2_K_S = 21
203208
LLAMA_FTYPE_GUESSED = 1024
204209

205210
# enum llama_rope_scaling_type {

vendor/llama.cpp

0 commit comments

Comments
 (0)