Add Cache flag to api (#1642)

This commit is contained in:
K0IN 2023-12-22 23:16:20 +01:00 committed by GitHub
parent db356c8519
commit 10da41d677
No known key found for this signature in database
GPG key ID: 4AEE18F83AFDEB23
3 changed files with 6 additions and 2 deletions

View file

@ -103,6 +103,7 @@ type Options struct {
MirostatEta float32 `json:"mirostat_eta,omitempty"`
PenalizeNewline bool `json:"penalize_newline,omitempty"`
Stop []string `json:"stop,omitempty"`
Cache bool `json:"cache,omitempty"`
}
// Runner options which must be set when the model is loaded into memory
@ -355,6 +356,7 @@ func DefaultOptions() Options {
MirostatEta: 0.1,
PenalizeNewline: true,
Seed: -1,
Cache: true,
Runner: Runner{
// options set when the model is loaded

View file

@ -306,7 +306,8 @@ curl http://localhost:11434/api/generate -d '{
"embedding_only": false,
"rope_frequency_base": 1.1,
"rope_frequency_scale": 0.8,
"num_thread": 8
"num_thread": 8,
"cache": true
}
}'
```

View file

@ -234,7 +234,8 @@ func predict(llm extServer, opts api.Options, ctx context.Context, predict Predi
"penalize_nl": opts.PenalizeNewline,
"seed": opts.Seed,
"stop": opts.Stop,
"image_data": imageData,
"image_data": imageData,
"cache_prompt": opts.Cache,
}
if predict.Format == "json" {