ollama/llm
Daniel Hiltgen d4cd695759 Add cgo implementation for llama.cpp
Run the server.cpp directly inside the Go runtime via cgo
while retaining the LLM Go abstractions.
2023-12-19 09:05:46 -08:00
..
llama.cpp Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
ext_server.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
ggml.go deprecate ggml 2023-12-19 09:05:46 -08:00
gguf.go remove per-model types 2023-12-11 09:40:21 -08:00
gpu_cuda.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
gpu_darwin.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
llama.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
llm.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00