llama.cpp/llama_cpp/server
Phil H 76aafa6149
Implement GGUF metadata KV overrides (#1011)
* Implement GGUF metadata overrides

* whitespace fix

* Fix kv overrides.

* Fix pointer and pickle

* Match llama.cpp kv_overrides cli argument

---------

Co-authored-by: Andrei <abetlen@gmail.com>
2024-01-15 12:29:29 -05:00
..
__init__.py llama_cpp server: app is now importable, still runnable as a module 2023-04-29 11:41:25 -07:00
__main__.py [Feat] Multi model support (#931) 2023-12-22 05:51:25 -05:00
app.py [Feat] Multi model support (#931) 2023-12-22 05:51:25 -05:00
cli.py [Feat] Multi model support (#931) 2023-12-22 05:51:25 -05:00
errors.py server: Support none defaulting to infinity for completions (#111) 2023-12-22 14:05:13 -05:00
model.py Implement GGUF metadata KV overrides (#1011) 2024-01-15 12:29:29 -05:00
settings.py Implement GGUF metadata KV overrides (#1011) 2024-01-15 12:29:29 -05:00
types.py server: Support none defaulting to infinity for completions (#111) 2023-12-22 14:05:13 -05:00