llama.cpp/llama_cpp
tc-wolf 4924455dec
feat: Make saved state more compact on-disk (#1296)
* State load/save changes

- Only store up to `n_tokens` logits instead of full `(n_ctx, n_vocab)`
  sized array.
  - Difference between ~350MB and ~1500MB for example prompt with ~300
    tokens (makes sense lol)
- Auto-formatting changes

* Back out formatting changes
2024-04-17 10:06:50 -04:00
..
server feat: Use all available CPUs for batch processing (#1345) 2024-04-17 10:05:54 -04:00
__init__.py chore: Bump version 2024-04-10 03:44:46 -04:00
_internals.py fix: last tokens passing to sample_repetition_penalties function (#1295) 2024-04-01 15:25:43 -04:00
_logger.py fix: Use llama_log_callback to avoid suppress_stdout_stderr 2024-02-05 21:52:12 -05:00
_utils.py Revert "Fix: fileno error google colab (#729) (#1156)" (#1157) 2024-02-02 12:18:55 -05:00
llama.py feat: Make saved state more compact on-disk (#1296) 2024-04-17 10:06:50 -04:00
llama_cache.py Move cache classes to llama_cache submodule. 2024-01-17 09:09:12 -05:00
llama_chat_format.py fix: pass correct type to chat handlers for chat completion logprobs 2024-04-10 03:41:55 -04:00
llama_cpp.py feat: Add typechecking for ctypes structure attributes 2024-04-10 02:40:41 -04:00
llama_grammar.py Fixed json strings grammar by blacklisting character control set. Closes #1259 2024-03-08 21:10:53 -05:00
llama_speculative.py Add speculative decoding (#1120) 2024-01-31 14:08:14 -05:00
llama_tokenizer.py fix: LlamaHFTokenizer now receives pre_tokens 2024-02-23 12:23:24 -05:00
llama_types.py feat: Add logprobs support to chat completions (#1311) 2024-03-31 13:30:13 -04:00
llava_cpp.py misc: llava_cpp use ctypes function decorator for binding 2024-02-26 11:07:33 -05:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00