Move workaround to new sample method
This commit is contained in:
parent
a4a1bbeaa9
commit
353e18a781
1 changed files with 5 additions and 5 deletions
|
@ -169,6 +169,11 @@ class Llama:
|
||||||
The sampled token.
|
The sampled token.
|
||||||
"""
|
"""
|
||||||
assert self.ctx is not None
|
assert self.ctx is not None
|
||||||
|
# Temporary workaround for https://github.com/ggerganov/llama.cpp/issues/684
|
||||||
|
if temp == 0.0:
|
||||||
|
temp = 1.0
|
||||||
|
top_p = 0.0
|
||||||
|
top_k = 1
|
||||||
return llama_cpp.llama_sample_top_p_top_k(
|
return llama_cpp.llama_sample_top_p_top_k(
|
||||||
ctx=self.ctx,
|
ctx=self.ctx,
|
||||||
last_n_tokens_data=(llama_cpp.llama_token * self.last_n_tokens_size)(
|
last_n_tokens_data=(llama_cpp.llama_token * self.last_n_tokens_size)(
|
||||||
|
@ -209,11 +214,6 @@ class Llama:
|
||||||
Yields:
|
Yields:
|
||||||
The generated tokens.
|
The generated tokens.
|
||||||
"""
|
"""
|
||||||
# Temporary workaround for https://github.com/ggerganov/llama.cpp/issues/684
|
|
||||||
if temp == 0.0:
|
|
||||||
temp = 1.0
|
|
||||||
top_p = 0.0
|
|
||||||
top_k = 1
|
|
||||||
assert self.ctx is not None
|
assert self.ctx is not None
|
||||||
self.reset()
|
self.reset()
|
||||||
while True:
|
while True:
|
||||||
|
|
Loading…
Reference in a new issue