llama.cpp/examples
Sigbjørn Skjæret dbcf64cf07
feat: Support SPM infill (#1492)
* Support SPM infill

* typo--

* one less layer of parenthesis necessary

* new required internals

* manually add bos/eos if model requires it

* add bos even when unknown

This is identical behaviour to llama.cpp

I guess any model that doesn't use BOS is recent enough to have the add_bos_token metadata.

* don't add bos/eos on non-infill pre-tokenized prompt

* add tokenizer hack to remove leading space in suffix

* I keep forgetting metadata are strings

* check if bos exists

* add example

* add cls/sep instead of bos/eos for WPM vocab

* simplify

* color-code filtered suffix

---------

Co-authored-by: Andrei Betlen <abetlen@gmail.com>
2024-06-13 03:45:24 -04:00
..
batch-processing feat: Add support for yaml based configs 2024-04-10 02:47:01 -04:00
gradio_chat docs(examples): Add gradio chat example 2024-02-25 21:09:13 -05:00
hf_pull docs(examples): Add huggingface pull example 2024-02-25 21:09:41 -05:00
high_level_api feat: Support SPM infill (#1492) 2024-06-13 03:45:24 -04:00
low_level_api examples: fix quantize example (#1387) 2024-04-27 20:48:47 -04:00
notebooks docs: Update multimodal example 2024-02-26 11:34:45 -05:00
ray example: LLM inference with Ray Serve (#1465) 2024-05-17 13:27:26 -04:00