Merge pull request #385 from nb-programmer/main
Update llama.py: Added how many input tokens in ValueError exception
This commit is contained in:
commit
ff9faaa48b
1 changed files with 1 additions and 1 deletions
|
@ -814,7 +814,7 @@ class Llama:
|
||||||
llama_cpp.llama_reset_timings(self.ctx)
|
llama_cpp.llama_reset_timings(self.ctx)
|
||||||
|
|
||||||
if len(prompt_tokens) > self._n_ctx:
|
if len(prompt_tokens) > self._n_ctx:
|
||||||
raise ValueError(f"Requested tokens exceed context window of {self._n_ctx}")
|
raise ValueError(f"Requested tokens ({len(prompt_tokens)}) exceed context window of {self._n_ctx}")
|
||||||
|
|
||||||
# Truncate max_tokens if requested tokens would exceed the context window
|
# Truncate max_tokens if requested tokens would exceed the context window
|
||||||
max_tokens = (
|
max_tokens = (
|
||||||
|
|
Loading…
Reference in a new issue