do not cache prompt (#2018)

- prompt cache causes inferance to hang after some time
This commit is contained in:
Bruce MacDonald 2024-01-16 13:48:05 -05:00 committed by GitHub
parent eef50accb4
commit a897e833b8
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -181,7 +181,6 @@ func (llm *dynExtServer) Predict(ctx context.Context, predict PredictOpts, fn fu
"seed": predict.Options.Seed,
"stop": predict.Options.Stop,
"image_data": imageData,
"cache_prompt": true,
}
if predict.Format == "json" {