ollama/llm
Bruce MacDonald 6fe178134d
improve api error handling (#781)
- remove new lines from llama.cpp error messages relayed to client
- check api option types and return error on wrong type
- change num layers from 95% VRAM to 92% VRAM
2023-10-13 16:57:10 -04:00
..
llama.cpp llm: fix build on amd64 2023-10-06 14:39:54 -07:00
falcon.go starcoder 2023-10-02 19:56:51 -07:00
ggml.go starcoder 2023-10-02 19:56:51 -07:00
gguf.go starcoder 2023-10-02 19:56:51 -07:00
llama.go improve api error handling (#781) 2023-10-13 16:57:10 -04:00
llm.go add format bytes 2023-10-11 14:08:23 -07:00
starcoder.go starcoder 2023-10-02 19:56:51 -07:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00