58d95cc9bd
This should resolve a number of memory leak and stability defects by allowing us to isolate llama.cpp in a separate process and shutdown when idle, and gracefully restart if it has problems. This also serves as a first step to be able to run multiple copies to support multiple models concurrently. |
||
---|---|---|
.. | ||
amd_common.go | ||
amd_hip_windows.go | ||
amd_linux.go | ||
amd_windows.go | ||
assets.go | ||
cpu_common.go | ||
gpu.go | ||
gpu_darwin.go | ||
gpu_info.h | ||
gpu_info_cpu.c | ||
gpu_info_cudart.c | ||
gpu_info_cudart.h | ||
gpu_info_darwin.h | ||
gpu_info_darwin.m | ||
gpu_info_nvml.c | ||
gpu_info_nvml.h | ||
gpu_test.go | ||
types.go |