do not cache prompt (#2018)
- prompt cache causes inferance to hang after some time
This commit is contained in:
parent
eef50accb4
commit
a897e833b8
1 changed files with 0 additions and 1 deletions
|
@ -181,7 +181,6 @@ func (llm *dynExtServer) Predict(ctx context.Context, predict PredictOpts, fn fu
|
|||
"seed": predict.Options.Seed,
|
||||
"stop": predict.Options.Stop,
|
||||
"image_data": imageData,
|
||||
"cache_prompt": true,
|
||||
}
|
||||
|
||||
if predict.Format == "json" {
|
||||
|
|
Loading…
Reference in a new issue