feat: Update llama.cpp
This commit is contained in:
parent
7981e9ce1e
commit
0653e15c20
2 changed files with 2 additions and 2 deletions
|
@ -2049,7 +2049,7 @@ llama_token_to_piece.restype = ctypes.c_int32
|
||||||
|
|
||||||
# /// Apply chat template. Inspired by hf apply_chat_template() on python.
|
# /// Apply chat template. Inspired by hf apply_chat_template() on python.
|
||||||
# /// Both "model" and "custom_template" are optional, but at least one is required. "custom_template" has higher precedence than "model"
|
# /// Both "model" and "custom_template" are optional, but at least one is required. "custom_template" has higher precedence than "model"
|
||||||
# /// NOTE: This function only support some known jinja templates. It is not a jinja parser.
|
# /// NOTE: This function does not use a jinja parser. It only support a pre-defined list of template. See more: https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template
|
||||||
# /// @param tmpl A Jinja template to use for this chat. If this is nullptr, the model’s default chat template will be used instead.
|
# /// @param tmpl A Jinja template to use for this chat. If this is nullptr, the model’s default chat template will be used instead.
|
||||||
# /// @param chat Pointer to a list of multiple llama_chat_message
|
# /// @param chat Pointer to a list of multiple llama_chat_message
|
||||||
# /// @param n_msg Number of llama_chat_message in this chat
|
# /// @param n_msg Number of llama_chat_message in this chat
|
||||||
|
|
2
vendor/llama.cpp
vendored
2
vendor/llama.cpp
vendored
|
@ -1 +1 @@
|
||||||
Subproject commit ba2135ccae7462470b3865c6e41d2e1d734eac05
|
Subproject commit 973053d8b0d04809836b3339a50f68d9c842de90
|
Loading…
Reference in a new issue