llama.cpp/llama_cpp/server
Jeffrey Fong 901827013b
feat: Integrate functionary v1.4 and v2 models + add custom tokenizer support to Llama class (#1078)
* convert functionary-v1 chat handler to use hf autotokenizer

* add hf_tokenizer + inteegrate functionary-v1.4 prompt template

* integrate functionary v2 prompt template

* update readme

* set up parallel function calling wip

* set up parallel function calling

* Update README.md

* Update README.md

* refactor tokenizers

* include old functionary handler for backward compatibility

* add hf_tokenizer_path in server ModelSettings

* convert functionary-v1 chat handler to use hf autotokenizer

* add hf_tokenizer + inteegrate functionary-v1.4 prompt template

* integrate functionary v2 prompt template

* update readme

* set up parallel function calling wip

* resolve merge conflict

* Update README.md

* Update README.md

* refactor tokenizers

* include old functionary handler for backward compatibility

* add hf_tokenizer_path in server ModelSettings

* Cleanup PR, fix breaking changes

* Use hf_pretrained_model_name_or_path for tokenizer

* fix hf tokenizer in streaming

* update README

* refactor offset mapping

---------

Co-authored-by: Andrei <abetlen@gmail.com>
2024-02-07 20:07:03 -05:00
..
__init__.py llama_cpp server: app is now importable, still runnable as a module 2023-04-29 11:41:25 -07:00
__main__.py [Feat] Multi model support (#931) 2023-12-22 05:51:25 -05:00
app.py feat(server): include llama-cpp-python version in openapi spec 2024-01-25 11:23:18 -05:00
cli.py Fix python3.8 support 2024-01-19 08:17:49 -05:00
errors.py server: Support none defaulting to infinity for completions (#111) 2023-12-22 14:05:13 -05:00
model.py feat: Integrate functionary v1.4 and v2 models + add custom tokenizer support to Llama class (#1078) 2024-02-07 20:07:03 -05:00
settings.py Add speculative decoding (#1120) 2024-01-31 14:08:14 -05:00
types.py server: Support none defaulting to infinity for completions (#111) 2023-12-22 14:05:13 -05:00