llama.cpp/examples/notebooks
Andrei ab028cb878
Migrate inference to llama_batch and llama_decode api (#795)
* Add low-level batching notebook

* fix: tokenization of special characters: (#850)

It should behave like llama.cpp, where most out of the box usages
treat special characters accordingly

* Update CHANGELOG

* Cleanup

* Fix runner label

* Update notebook

* Use llama_decode and batch api

* Support logits_all parameter

---------

Co-authored-by: Antoine Lizee <antoine.lizee@gmail.com>
2023-11-02 20:13:57 -04:00
..
Batching.ipynb Migrate inference to llama_batch and llama_decode api (#795) 2023-11-02 20:13:57 -04:00
Clients.ipynb Add clients example. Closes #46 2023-04-08 09:35:32 -04:00
Guidance.ipynb Add Guidance example 2023-05-19 03:16:58 -04:00
PerformanceTuning.ipynb Add performance tuning notebook 2023-04-05 04:09:19 -04:00