ollama/llm
Bruce MacDonald bbe41ce41a
fix: parallel queueing race condition caused silent failure (#1445)
* fix: queued request failures

- increase parallel requests to 2 to complete queued request, queueing is managed in ollama

* log steam errors
2023-12-09 14:14:02 -05:00
..
llama.cpp update for qwen 2023-12-04 11:38:05 -08:00
falcon.go starcoder 2023-10-02 19:56:51 -07:00
ggml.go seek instead of copyn 2023-12-04 16:59:23 -08:00
gguf.go comments 2023-12-04 16:59:23 -08:00
llama.go fix: parallel queueing race condition caused silent failure (#1445) 2023-12-09 14:14:02 -05:00
llm.go load projectors 2023-12-05 14:36:12 -08:00
starcoder.go starcoder 2023-10-02 19:56:51 -07:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00