Move workaround to new sample method

This commit is contained in:
Andrei Betlen 2023-04-02 00:06:34 -04:00
parent a4a1bbeaa9
commit 353e18a781

View file

@ -169,6 +169,11 @@ class Llama:
The sampled token. The sampled token.
""" """
assert self.ctx is not None assert self.ctx is not None
# Temporary workaround for https://github.com/ggerganov/llama.cpp/issues/684
if temp == 0.0:
temp = 1.0
top_p = 0.0
top_k = 1
return llama_cpp.llama_sample_top_p_top_k( return llama_cpp.llama_sample_top_p_top_k(
ctx=self.ctx, ctx=self.ctx,
last_n_tokens_data=(llama_cpp.llama_token * self.last_n_tokens_size)( last_n_tokens_data=(llama_cpp.llama_token * self.last_n_tokens_size)(
@ -209,11 +214,6 @@ class Llama:
Yields: Yields:
The generated tokens. The generated tokens.
""" """
# Temporary workaround for https://github.com/ggerganov/llama.cpp/issues/684
if temp == 0.0:
temp = 1.0
top_p = 0.0
top_k = 1
assert self.ctx is not None assert self.ctx is not None
self.reset() self.reset()
while True: while True: