Jeffrey Morgan
c63f811909
return error if model fails to load
2023-07-11 20:32:26 -07:00
Jeffrey Morgan
7c71c10d4f
fix compilation issue in Dockerfile, remove from README.md
until ready
2023-07-11 19:51:08 -07:00
Jeffrey Morgan
e64ef69e34
look for ggml-metal in the same directory as the binary
2023-07-11 15:58:56 -07:00
Michael Yang
442dec1c6f
vendor llama.cpp
2023-07-11 11:59:18 -07:00
Michael Yang
fd4792ec56
call llama.cpp directly from go
2023-07-11 11:59:18 -07:00
Jeffrey Morgan
268e362fa7
fix binding build
2023-07-10 11:33:43 -07:00
Jeffrey Morgan
a18e6b3a40
llama: remove unnecessary std::vector
2023-07-09 10:51:45 -04:00
Jeffrey Morgan
5fb96255dc
llama: remove unused helper functions
2023-07-09 10:25:07 -04:00
Patrick Devine
3f1b7177f2
pass model and predict options
2023-07-07 09:34:05 -07:00
Michael Yang
5dc9c8ff23
more free
2023-07-06 17:08:03 -07:00
Bruce MacDonald
da74384a3e
remove prompt cache
2023-07-06 17:49:05 -04:00
Michael Yang
2c80eddd71
more free
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
9fe018675f
use Makefile
for dependency building instead of go generate
2023-07-06 16:34:44 -04:00
Michael Yang
1b7183c5a1
enable metal gpu acceleration
...
ggml-metal.metal must be in the same directory as the ollama binary
otherwise llama.cpp will not be able to find it and load it.
1. go generate llama/llama_metal.go
2. go build .
3. ./ollama serve
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
0998d4f0a4
remove debug print statements
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
79a999e95d
fix crash in bindings
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
fd962a36e5
client updates
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
0240165388
fix llama.cpp build
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
9164981d72
move prompt templates out of python bindings
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
6093a88c1a
add llama.cpp go bindings
2023-07-06 16:34:44 -04:00