llama.cpp/examples/notebooks
Damian Stewart aab74f0b2b
Multimodal Support (Llava 1.5) (#821)
* llava v1.5 integration

* Point llama.cpp to fork

* Add llava shared library target

* Fix type

* Update llama.cpp

* Add llava api

* Revert changes to llama and llama_cpp

* Update llava example

* Add types for new gpt-4-vision-preview api

* Fix typo

* Update llama.cpp

* Update llama_types to match OpenAI v1 API

* Update ChatCompletionFunction type

* Reorder request parameters

* More API type fixes

* Even More Type Updates

* Add parameter for custom chat_handler to Llama class

* Fix circular import

* Convert to absolute imports

* Fix

* Fix pydantic Jsontype bug

* Accept list of prompt tokens in create_completion

* Add llava1.5 chat handler

* Add Multimodal notebook

* Clean up examples

* Add server docs

---------

Co-authored-by: Andrei Betlen <abetlen@gmail.com>
2023-11-07 22:48:51 -05:00
..
Batching.ipynb Migrate inference to llama_batch and llama_decode api (#795) 2023-11-02 20:13:57 -04:00
Clients.ipynb Add clients example. Closes #46 2023-04-08 09:35:32 -04:00
Functions.ipynb Add functionary support (#784) 2023-11-03 02:12:14 -04:00
Guidance.ipynb Add Guidance example 2023-05-19 03:16:58 -04:00
Multimodal.ipynb Multimodal Support (Llava 1.5) (#821) 2023-11-07 22:48:51 -05:00
PerformanceTuning.ipynb Add performance tuning notebook 2023-04-05 04:09:19 -04:00