ollama/llm
Bruce MacDonald 2540c9181c
support for packaging in multiple cuda runners (#509)
* enable packaging multiple cuda versions
* use nvcc cuda version if available

---------

Co-authored-by: Michael Yang <mxyng@pm.me>
2023-09-14 15:08:13 -04:00
..
llama.cpp support for packaging in multiple cuda runners (#509) 2023-09-14 15:08:13 -04:00
falcon.go fix: add falcon.go 2023-09-13 14:47:37 -07:00
ggml.go support for packaging in multiple cuda runners (#509) 2023-09-14 15:08:13 -04:00
gguf.go support for packaging in multiple cuda runners (#509) 2023-09-14 15:08:13 -04:00
llama.go support for packaging in multiple cuda runners (#509) 2023-09-14 15:08:13 -04:00
llm.go fix falcon decode 2023-09-12 12:34:53 -07:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00