Bruce MacDonald
42998d797d
subprocess llama.cpp server ( #401 )
...
* remove c code
* pack llama.cpp
* use request context for llama_cpp
* let llama_cpp decide the number of threads to use
* stop llama runner when app stops
* remove sample count and duration metrics
* use go generate to get libraries
* tmp dir for running llm
2023-08-30 16:35:03 -04:00
Jeffrey Morgan
67b6f8ba86
add ggml-metal.metal
to .gitignore
2023-07-28 11:04:21 -04:00
jk1jk
e6c427ce4d
Update .gitignore
2023-07-22 17:00:52 +03:00
Jeffrey Morgan
7c71c10d4f
fix compilation issue in Dockerfile, remove from README.md
until ready
2023-07-11 19:51:08 -07:00
Michael Yang
442dec1c6f
vendor llama.cpp
2023-07-11 11:59:18 -07:00
Michael Yang
fd4792ec56
call llama.cpp directly from go
2023-07-11 11:59:18 -07:00
Jeffrey Morgan
9fe018675f
use Makefile
for dependency building instead of go generate
2023-07-06 16:34:44 -04:00
Jeffrey Morgan
b0e986fb96
add binary to .gitignore
2023-07-06 16:34:44 -04:00
Bruce MacDonald
d34985b9df
add templates to prompt command
2023-06-26 13:41:16 -04:00
Jeffrey Morgan
b361fa72ec
reorganize directories
2023-06-25 13:08:03 -04:00
Jeffrey Morgan
d3709f85b5
build server into desktop app
2023-06-25 00:30:02 -04:00
Bruce MacDonald
c5bafaff54
package server with client
2023-06-23 18:38:22 -04:00
Bruce MacDonald
f0eee3faa0
build server executable
2023-06-23 17:23:30 -04:00
Bruce MacDonald
db81d81b23
Update .gitignore
2023-06-23 13:57:03 -04:00
Jeffrey Morgan
8fa91332fa
initial commit
2023-06-22 18:31:40 -04:00