58d95cc9bd
This should resolve a number of memory leak and stability defects by allowing us to isolate llama.cpp in a separate process and shutdown when idle, and gracefully restart if it has problems. This also serves as a first step to be able to run multiple copies to support multiple models concurrently.
6 lines
84 B
Go
6 lines
84 B
Go
package llm
|
|
|
|
import "embed"
|
|
|
|
//go:embed build/linux/*/*/bin/*
|
|
var libEmbed embed.FS
|