From 0653e15c206b5ef1f8cb064d2cd18beb662e6cd1 Mon Sep 17 00:00:00 2001 From: Andrei Betlen Date: Wed, 21 Feb 2024 23:04:52 -0500 Subject: [PATCH] feat: Update llama.cpp --- llama_cpp/llama_cpp.py | 2 +- vendor/llama.cpp | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/llama_cpp/llama_cpp.py b/llama_cpp/llama_cpp.py index 69dbe09..2ab326d 100644 --- a/llama_cpp/llama_cpp.py +++ b/llama_cpp/llama_cpp.py @@ -2049,7 +2049,7 @@ llama_token_to_piece.restype = ctypes.c_int32 # /// Apply chat template. Inspired by hf apply_chat_template() on python. # /// Both "model" and "custom_template" are optional, but at least one is required. "custom_template" has higher precedence than "model" -# /// NOTE: This function only support some known jinja templates. It is not a jinja parser. +# /// NOTE: This function does not use a jinja parser. It only support a pre-defined list of template. See more: https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template # /// @param tmpl A Jinja template to use for this chat. If this is nullptr, the model’s default chat template will be used instead. # /// @param chat Pointer to a list of multiple llama_chat_message # /// @param n_msg Number of llama_chat_message in this chat diff --git a/vendor/llama.cpp b/vendor/llama.cpp index ba2135c..973053d 160000 --- a/vendor/llama.cpp +++ b/vendor/llama.cpp @@ -1 +1 @@ -Subproject commit ba2135ccae7462470b3865c6e41d2e1d734eac05 +Subproject commit 973053d8b0d04809836b3339a50f68d9c842de90