Update llama.cpp

This commit is contained in:
Andrei Betlen 2023-05-07 00:12:47 -04:00
parent bc853e3742
commit 7c3743fe5f
3 changed files with 9 additions and 5 deletions

View file

@ -297,21 +297,25 @@ class Llama:
ctx=self.ctx, ctx=self.ctx,
candidates=llama_cpp.ctypes.pointer(candidates), candidates=llama_cpp.ctypes.pointer(candidates),
k=top_k, k=top_k,
min_keep=llama_cpp.c_size_t(1),
) )
llama_cpp.llama_sample_tail_free( llama_cpp.llama_sample_tail_free(
ctx=self.ctx, ctx=self.ctx,
candidates=llama_cpp.ctypes.pointer(candidates), candidates=llama_cpp.ctypes.pointer(candidates),
z=llama_cpp.c_float(1.0), z=llama_cpp.c_float(1.0),
min_keep=llama_cpp.c_size_t(1),
) )
llama_cpp.llama_sample_typical( llama_cpp.llama_sample_typical(
ctx=self.ctx, ctx=self.ctx,
candidates=llama_cpp.ctypes.pointer(candidates), candidates=llama_cpp.ctypes.pointer(candidates),
p=llama_cpp.c_float(1.0), p=llama_cpp.c_float(1.0),
min_keep=llama_cpp.c_size_t(1),
) )
llama_cpp.llama_sample_top_p( llama_cpp.llama_sample_top_p(
ctx=self.ctx, ctx=self.ctx,
candidates=llama_cpp.ctypes.pointer(candidates), candidates=llama_cpp.ctypes.pointer(candidates),
p=top_p, p=top_p,
min_keep=llama_cpp.c_size_t(1),
) )
llama_cpp.llama_sample_temperature( llama_cpp.llama_sample_temperature(
ctx=self.ctx, ctx=self.ctx,

View file

@ -515,7 +515,7 @@ def llama_sample_top_k(
ctx: llama_context_p, ctx: llama_context_p,
candidates, # type: _Pointer[llama_token_data_array] candidates, # type: _Pointer[llama_token_data_array]
k: c_int, k: c_int,
min_keep: c_size_t = c_size_t(1), min_keep: c_size_t,
): ):
return _lib.llama_sample_top_k(ctx, candidates, k, min_keep) return _lib.llama_sample_top_k(ctx, candidates, k, min_keep)
@ -534,7 +534,7 @@ def llama_sample_top_p(
ctx: llama_context_p, ctx: llama_context_p,
candidates, # type: _Pointer[llama_token_data_array] candidates, # type: _Pointer[llama_token_data_array]
p: c_float, p: c_float,
min_keep: c_size_t = c_size_t(1), min_keep: c_size_t,
): ):
return _lib.llama_sample_top_p(ctx, candidates, p, min_keep) return _lib.llama_sample_top_p(ctx, candidates, p, min_keep)
@ -553,7 +553,7 @@ def llama_sample_tail_free(
ctx: llama_context_p, ctx: llama_context_p,
candidates, # type: _Pointer[llama_token_data_array] candidates, # type: _Pointer[llama_token_data_array]
z: c_float, z: c_float,
min_keep: c_size_t = c_size_t(1), min_keep: c_size_t,
): ):
return _lib.llama_sample_tail_free(ctx, candidates, z, min_keep) return _lib.llama_sample_tail_free(ctx, candidates, z, min_keep)
@ -572,7 +572,7 @@ def llama_sample_typical(
ctx: llama_context_p, ctx: llama_context_p,
candidates, # type: _Pointer[llama_token_data_array] candidates, # type: _Pointer[llama_token_data_array]
p: c_float, p: c_float,
min_keep: c_size_t = c_size_t(1), min_keep: c_size_t,
): ):
return _lib.llama_sample_typical(ctx, candidates, p, min_keep) return _lib.llama_sample_typical(ctx, candidates, p, min_keep)

2
vendor/llama.cpp vendored

@ -1 +1 @@
Subproject commit 2edbdb0f99336cb41f0995061c7602ed54beb863 Subproject commit 1b0fd454650ef4d68a980e3225488b79e6e9af25