Commit graph

33 commits

Author SHA1 Message Date
bretello
740f3f3812
fix: set LLAMA_METAL_EMBED_LIBRARY=on on MacOS arm64 (#1289) 2024-03-20 12:46:09 -04:00
Douglas Hanley
19b55ad3e5
feat: use gpu backend for clip if available (#1175) 2024-02-11 13:53:59 -05:00
Andrei Betlen
e3941d9c67 Make building llava optional 2023-11-28 04:55:21 -05:00
Andrei Betlen
4388f33414 Set CUDA_ARCHITECTURES=OFF for windows 2023-11-10 16:32:36 -05:00
Damian Stewart
aab74f0b2b
Multimodal Support (Llava 1.5) (#821)
* llava v1.5 integration

* Point llama.cpp to fork

* Add llava shared library target

* Fix type

* Update llama.cpp

* Add llava api

* Revert changes to llama and llama_cpp

* Update llava example

* Add types for new gpt-4-vision-preview api

* Fix typo

* Update llama.cpp

* Update llama_types to match OpenAI v1 API

* Update ChatCompletionFunction type

* Reorder request parameters

* More API type fixes

* Even More Type Updates

* Add parameter for custom chat_handler to Llama class

* Fix circular import

* Convert to absolute imports

* Fix

* Fix pydantic Jsontype bug

* Accept list of prompt tokens in create_completion

* Add llava1.5 chat handler

* Add Multimodal notebook

* Clean up examples

* Add server docs

---------

Co-authored-by: Andrei Betlen <abetlen@gmail.com>
2023-11-07 22:48:51 -05:00
Andrei Betlen
b4939c2d99 Revert BUILD_NUMBER fix 2023-09-29 23:28:45 -04:00
Andrei Betlen
43f8fc371a Potential fix for pip install bug 2023-09-29 22:24:22 -04:00
Andrei Betlen
8d75016549 Install required runtime dlls to package directory on windows 2023-09-16 14:57:49 -04:00
Andrei Betlen
dbd3a6d1ed Fix issue installing on m1 macs 2023-09-14 18:25:44 -04:00
Andrei Betlen
e3387e4655 Add explanatory comment 2023-09-12 18:41:00 -04:00
Andrei Betlen
e65a823dc8 Update flags 2023-09-12 18:37:32 -04:00
Andrei Betlen
4cb0e35b64 string options 2023-09-12 18:35:34 -04:00
Andrei Betlen
2c3df160fb Reorder 2023-09-12 18:29:18 -04:00
Andrei Betlen
d123129e63 fix 2023-09-12 18:25:31 -04:00
Andrei Betlen
dadfd96745 Use compiler to determine best optimizations for platform 2023-09-12 18:21:49 -04:00
Andrei Betlen
4c0787b408 Disable acceleration in macos tests only 2023-09-12 18:05:44 -04:00
Andrei Betlen
d24383eaef Disable acceleration on macos 2023-09-12 17:58:47 -04:00
Andrei Betlen
9547a351ee Try arm64 python 2023-09-12 17:35:07 -04:00
Andrei Betlen
010a50168c Add tune 2023-09-12 17:20:57 -04:00
Andrei Betlen
f93fb30f5c Set native arch flags for macos 2023-09-12 17:15:07 -04:00
Andrei Betlen
6cb77a20c6 Migrate to scikit-build-core. Closes #489 2023-07-18 18:52:29 -04:00
Andrei Betlen
c1eaef329a Add resource destination to cmake 2023-06-10 18:11:48 -04:00
Andrei Betlen
e3542b6627 Revert "Merge pull request #350 from abetlen/migrate-to-scikit-build-core"
This reverts commit fb2c5f7fd9, reversing
changes made to 202ed4464b.
2023-06-09 23:23:16 -04:00
Andrei Betlen
b025a859ae Add full path to shared library installation path 2023-06-08 22:11:01 -04:00
Andrei Betlen
7345456779 Migrate to scikit-build-core 2023-06-08 21:49:42 -04:00
Andrei Betlen
cf6a9d6d8e Add framework and archive destinations to cmake for macos and windows. 2023-06-07 03:42:55 -04:00
Andrei Betlen
341c50b5b0 Fix CMakeLists.txt 2023-05-16 09:07:14 -04:00
Andrei Betlen
848c83dfd0 Add FORCE_CMAKE option 2023-04-25 01:36:37 -04:00
Andrei Betlen
e611cfc56d Build shared library with make on unix platforms 2023-04-08 02:39:17 -04:00
Andrei Betlen
88c23d04a8 Fix windows dll location issue 2023-04-06 22:44:31 -04:00
Andrei Betlen
241722c981 Quote destination 2023-04-06 22:38:53 -04:00
Andrei Betlen
02760f8fa7 Update llama.cpp and shared library build process 2023-03-23 17:01:06 -04:00
Andrei Betlen
465238b179 Updated package to build with skbuild 2023-03-23 13:54:14 -04:00