ollama/llm
Bruce MacDonald 1fbf3585d6
Relay default values to llama runner (#672)
* include seed in params for llama.cpp server and remove empty filter for temp

* relay default predict options to llama.cpp

- reorganize options to match predict request for readability

* omit empty stop

---------

Co-authored-by: hallh <hallh@users.noreply.github.com>
2023-10-02 14:53:16 -04:00
..
llama.cpp silence warm up log 2023-09-21 14:53:33 -07:00
falcon.go fix: add falcon.go 2023-09-13 14:47:37 -07:00
ggml.go unbound max num gpu layers (#591) 2023-09-25 18:36:46 -04:00
gguf.go unbound max num gpu layers (#591) 2023-09-25 18:36:46 -04:00
llama.go Relay default values to llama runner (#672) 2023-10-02 14:53:16 -04:00
llm.go unbound max num gpu layers (#591) 2023-09-25 18:36:46 -04:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00