llama.cpp/llama_cpp
Connor a05d90446f
fix: Circular dependancy preventing early Llama object free (#1176)
commit 901827013b introduced a cyclic dependency
within Llama objects. That change causes old models to linger in memory longer
than necessary, thereby creating memory bloat in most applications attempting
to switch between models at runtime. This patch simply removes the problematic
line, allowing models to deallocate without relying on GC. One might also
consider combining `weakref.ref` with a `@property` if the `llama` attribute is
absolutely necessary to expose in the tokenizer class.
2024-02-11 13:57:57 -05:00
..
server fix: broken import 2024-02-08 01:13:28 -05:00
__init__.py Bump version 2024-02-06 12:47:59 -05:00
_internals.py fix: Use llama_log_callback to avoid suppress_stdout_stderr 2024-02-05 21:52:12 -05:00
_logger.py fix: Use llama_log_callback to avoid suppress_stdout_stderr 2024-02-05 21:52:12 -05:00
_utils.py Revert "Fix: fileno error google colab (#729) (#1156)" (#1157) 2024-02-02 12:18:55 -05:00
llama.py Merge branch 'main' of github.com:abetlen/llama_cpp_python into main 2024-02-09 13:32:31 -05:00
llama_cache.py Move cache classes to llama_cache submodule. 2024-01-17 09:09:12 -05:00
llama_chat_format.py feat: Integrate functionary v1.4 and v2 models + add custom tokenizer support to Llama class (#1078) 2024-02-07 20:07:03 -05:00
llama_cpp.py Update llama.cpp 2024-02-05 13:26:50 -05:00
llama_grammar.py fix: from_json_schema oneof/anyof bug. Closes #1097 2024-01-21 19:06:53 -05:00
llama_speculative.py Add speculative decoding (#1120) 2024-01-31 14:08:14 -05:00
llama_tokenizer.py fix: Circular dependancy preventing early Llama object free (#1176) 2024-02-11 13:57:57 -05:00
llama_types.py Add json schema mode (#1122) 2024-01-27 16:52:18 -05:00
llava_cpp.py Make building llava optional 2023-11-28 04:55:21 -05:00
py.typed Add py.typed 2023-08-11 09:58:48 +02:00