ollama/llm/llama.cpp
Daniel Hiltgen 9a70aecccb Refactor how we augment llama.cpp
This changes the model for llama.cpp inclusion so we're not applying a patch,
but instead have the C++ code directly in the ollama tree, which should make it
easier to refine and update over time.
2024-01-02 15:35:55 -08:00
..
gguf@328b83de23 Bump llama.cpp to b1662 and set n_parallel=1 2023-12-19 09:05:46 -08:00
ext_server.cpp Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00
ext_server.h Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00
gen_common.sh Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00
gen_darwin.sh Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00
gen_linux.sh Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00
gen_windows.ps1 Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00
generate_darwin.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
generate_linux.go Adapted rocm support to cgo based llama.cpp 2023-12-19 09:05:46 -08:00
generate_windows.go Add cgo implementation for llama.cpp 2023-12-19 09:05:46 -08:00
ollama.txt Refactor how we augment llama.cpp 2024-01-02 15:35:55 -08:00