Bugfix: only eval new tokens

This commit is contained in:
Andrei Betlen 2023-04-15 17:32:53 -04:00
parent 887f3b73ac
commit 89856ef00d

View file

@ -280,6 +280,7 @@ class Llama:
if self.verbose: if self.verbose:
print("generate cache hit", file=sys.stderr) print("generate cache hit", file=sys.stderr)
reset = False reset = False
tokens = tokens[len(self.tokens) :]
### ###
if reset: if reset:
self.reset() self.reset()