Commit graph

62 commits

Author SHA1 Message Date
Patrick Devine
3f1b7177f2 pass model and predict options 2023-07-07 09:34:05 -07:00
Michael Yang
5dc9c8ff23 more free 2023-07-06 17:08:03 -07:00
Bruce MacDonald
da74384a3e remove prompt cache 2023-07-06 17:49:05 -04:00
Michael Yang
2c80eddd71 more free 2023-07-06 16:34:44 -04:00
Jeffrey Morgan
9fe018675f use Makefile for dependency building instead of go generate 2023-07-06 16:34:44 -04:00
Michael Yang
1b7183c5a1 enable metal gpu acceleration
ggml-metal.metal must be in the same directory as the ollama binary
otherwise llama.cpp will not be able to find it and load it.

1. go generate llama/llama_metal.go
2. go build .
3. ./ollama serve
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
0998d4f0a4 remove debug print statements 2023-07-06 16:34:44 -04:00
Jeffrey Morgan
79a999e95d fix crash in bindings 2023-07-06 16:34:44 -04:00
Jeffrey Morgan
fd962a36e5 client updates 2023-07-06 16:34:44 -04:00
Jeffrey Morgan
0240165388 fix llama.cpp build 2023-07-06 16:34:44 -04:00
Jeffrey Morgan
9164981d72 move prompt templates out of python bindings 2023-07-06 16:34:44 -04:00
Jeffrey Morgan
6093a88c1a add llama.cpp go bindings 2023-07-06 16:34:44 -04:00