llama.cpp/llama_cpp/server
Felipe Lorenz c139f8b5d5
feat: Add endpoints for tokenize, detokenize and count tokens (#1136)
* Add endpoint to count tokens

* Add tokenize and detokenize endpoints

* Change response key to tokens for tokenize endpoint

* Fix dependency bug

* Cleanup

* Remove example added by mistake

* Move tokenize, detokenize, and count to Extras namespace. Tag existing endpoints

---------

Co-authored-by: Andrei Betlen <abetlen@gmail.com>
2024-03-08 21:09:00 -05:00
..
__init__.py llama_cpp server: app is now importable, still runnable as a module 2023-04-29 11:41:25 -07:00
__main__.py [Feat] Multi model support (#931) 2023-12-22 05:51:25 -05:00
app.py feat: Add endpoints for tokenize, detokenize and count tokens (#1136) 2024-03-08 21:09:00 -05:00
cli.py Fix python3.8 support 2024-01-19 08:17:49 -05:00
errors.py misc: Format 2024-02-28 14:27:40 -05:00
model.py misc: Format 2024-02-28 14:27:40 -05:00
settings.py misc: Format 2024-02-28 14:27:40 -05:00
types.py feat: Add endpoints for tokenize, detokenize and count tokens (#1136) 2024-03-08 21:09:00 -05:00