llama.cpp/llama_cpp
Phil H fe5d6ea648
fix: GGUF metadata KV overrides, re #1011 (#1116)
* kv overrides another attempt

* add sentinel element, simplify array population

* ensure sentinel element is zeroed
2024-01-23 22:00:38 -05:00
..
server fix: pass chat handler not chat formatter for huggingface autotokenizer and tokenizer_config formats. 2024-01-21 18:38:04 -05:00
__init__.py Bump version 2024-01-21 20:30:24 -05:00
_internals.py feat: Expose gguf model metadata in metadata property 2024-01-19 10:46:03 -05:00
_utils.py feat: Add ability to load chat format from huggingface autotokenizer or tokenizer_config.json files. 2024-01-18 21:21:37 -05:00
llama.py fix: GGUF metadata KV overrides, re #1011 (#1116) 2024-01-23 22:00:38 -05:00
llama_cache.py Move cache classes to llama_cache submodule. 2024-01-17 09:09:12 -05:00
llama_chat_format.py fix: use both eos and bos tokens as stop sequences for hf-tokenizer-config chat format. 2024-01-22 08:32:48 -05:00
llama_cpp.py Update llama.cpp 2024-01-22 11:25:11 -05:00
llama_grammar.py fix: from_json_schema oneof/anyof bug. Closes #1097 2024-01-21 19:06:53 -05:00
llama_types.py Add missing tool_calls finish_reason 2023-11-10 02:51:06 -05:00
llava_cpp.py Make building llava optional 2023-11-28 04:55:21 -05:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00