Bruce MacDonald
42998d797d
subprocess llama.cpp server ( #401 )
...
* remove c code
* pack llama.cpp
* use request context for llama_cpp
* let llama_cpp decide the number of threads to use
* stop llama runner when app stops
* remove sample count and duration metrics
* use go generate to get libraries
* tmp dir for running llm
2023-08-30 16:35:03 -04:00
Michael Yang
041f9ad1a1
update README.md
2023-08-25 11:44:25 -07:00
Jeffrey Morgan
1f78e409b4
docs: format with prettier
2023-08-08 15:41:48 -07:00
Michael Yang
24e43e3212
update development.md
2023-07-24 09:43:57 -07:00
Bruce MacDonald
52f04e39f2
Note that CGO must be enabled in dev docs
2023-07-21 22:36:36 +02:00
Matt Williams
3d9498dc95
Some simple modelfile examples
...
Signed-off-by: Matt Williams <m@technovangelist.com>
2023-07-17 17:16:59 -07:00
Jeffrey Morgan
1358e27b77
add publish script
2023-07-07 12:59:45 -04:00
Michael Yang
9811956938
update development.md
2023-06-28 12:41:30 -07:00
Jeffrey Morgan
9ba58c8a9e
move desktop docs to desktop/
2023-06-28 11:29:29 -04:00
Jeffrey Morgan
9f868d8258
move desktop docs to desktop/
2023-06-28 11:27:18 -04:00
Bruce MacDonald
4018b3c533
poetry development
2023-06-28 11:17:08 -04:00
Bruce MacDonald
ecfb4abafb
simplify loading
2023-06-27 14:50:30 -04:00
Michael Chiang
2906cbab11
Update development.md
2023-06-27 14:07:31 -04:00
Michael Chiang
9d14e75185
Update development.md
2023-06-27 14:06:59 -04:00
Michael Chiang
a2745f8174
Update development.md
2023-06-27 14:06:49 -04:00
Jeffrey Morgan
20cdd9fee6
update README.md
2023-06-27 13:51:20 -04:00
Bruce MacDonald
11614b6d84
add development doc
2023-06-27 13:46:46 -04:00