diff --git a/llama_cpp/server/model.py b/llama_cpp/server/model.py index c2d6b6d..bbb6806 100644 --- a/llama_cpp/server/model.py +++ b/llama_cpp/server/model.py @@ -78,7 +78,7 @@ class LlamaProxy: settings.hf_pretrained_model_name_or_path is not None ), "hf_pretrained_model_name_or_path must be set for hf-autotokenizer" chat_handler = ( - llama_cpp.llama_chat_format.hf_autotokenizer_to_chat_formatter( + llama_cpp.llama_chat_format.hf_autotokenizer_to_chat_completion_handler( settings.hf_pretrained_model_name_or_path ) ) @@ -87,7 +87,7 @@ class LlamaProxy: settings.hf_tokenizer_config_path is not None ), "hf_tokenizer_config_path must be set for hf-tokenizer-config" chat_handler = ( - llama_cpp.llama_chat_format.hf_tokenizer_config_to_chat_formatter( + llama_cpp.llama_chat_format.hf_tokenizer_config_to_chat_completion_handler( json.load(open(settings.hf_tokenizer_config_path)) ) )