diff --git a/llama_cpp/llama.py b/llama_cpp/llama.py index b43dfe7..0978e1e 100644 --- a/llama_cpp/llama.py +++ b/llama_cpp/llama.py @@ -696,9 +696,7 @@ class Llama: llama_cpp.llama_reset_timings(self.ctx) if len(prompt_tokens) + max_tokens > self._n_ctx: - raise ValueError( - f"Requested tokens exceed context window of {self._n_ctx}" - ) + raise ValueError(f"Requested tokens exceed context window of {self._n_ctx}") if stop != []: stop_sequences = [s.encode("utf-8") for s in stop]