llama.cpp/llama_cpp
2024-01-19 09:03:35 -05:00
..
server Fix python3.8 support 2024-01-19 08:17:49 -05:00
__init__.py Bump version 2024-01-19 09:03:35 -05:00
_internals.py Move helper classes to _internals submodule 2024-01-17 09:14:00 -05:00
_utils.py feat: Add ability to load chat format from huggingface autotokenizer or tokenizer_config.json files. 2024-01-18 21:21:37 -05:00
llama.py Fix mirostat sampling 2024-01-19 08:31:59 -05:00
llama_cache.py Move cache classes to llama_cache submodule. 2024-01-17 09:09:12 -05:00
llama_chat_format.py feat: Add ability to load chat format from huggingface autotokenizer or tokenizer_config.json files. 2024-01-18 21:21:37 -05:00
llama_cpp.py Update llama.cpp 2024-01-18 21:21:49 -05:00
llama_grammar.py Fix Pydantic model parsing (#1087) 2024-01-15 10:45:57 -05:00
llama_types.py Add missing tool_calls finish_reason 2023-11-10 02:51:06 -05:00
llava_cpp.py Make building llava optional 2023-11-28 04:55:21 -05:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00