ollama/llm
Daniel Hiltgen e1f50377f4 Harden generate patching model
Only apply patches if we have any, and make sure to cleanup
every file we patched at the end to leave the tree clean
2024-02-01 19:34:36 -08:00
..
ext_server Bump llama.cpp to b1999 2024-01-30 16:52:12 -08:00
generate Harden generate patching model 2024-02-01 19:34:36 -08:00
llama.cpp@d2f650cb5b Bump llama.cpp to b1999 2024-01-30 16:52:12 -08:00
patches Bump llama.cpp to b1999 2024-01-30 16:52:12 -08:00
dyn_ext_server.c Switch to local dlopen symbols 2024-01-19 11:37:02 -08:00
dyn_ext_server.go use llm.ImageData 2024-01-31 19:13:48 -08:00
dyn_ext_server.h Always dynamically load the llm server library 2024-01-11 08:42:47 -08:00
ggml.go add max context length check 2024-01-12 14:54:07 -08:00
gguf.go refactor tensor read 2024-01-24 10:48:31 -08:00
llama.go use llm.ImageData 2024-01-31 19:13:48 -08:00
llm.go Load all layers on arm64 macOS if model is small enough (#2149) 2024-01-22 17:40:06 -08:00
payload_common.go use gzip for runner embedding (#2067) 2024-01-19 13:23:03 -05:00
payload_darwin_amd64.go Add multiple CPU variants for Intel Mac 2024-01-17 15:08:54 -08:00
payload_darwin_arm64.go Add multiple CPU variants for Intel Mac 2024-01-17 15:08:54 -08:00
payload_linux.go Add multiple CPU variants for Intel Mac 2024-01-17 15:08:54 -08:00
payload_test.go Fix up the CPU fallback selection 2024-01-11 15:27:06 -08:00
payload_windows.go Add multiple CPU variants for Intel Mac 2024-01-17 15:08:54 -08:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00