2023-12-23 19:35:44 +00:00
|
|
|
package llm
|
|
|
|
|
|
|
|
import (
|
|
|
|
"github.com/jmorganca/ollama/api"
|
|
|
|
)
|
|
|
|
|
|
|
|
func newDefaultExtServer(model string, adapters, projectors []string, numLayers int64, opts api.Options) (extServer, error) {
|
|
|
|
// On windows we always load the llama.cpp libraries dynamically to avoid startup DLL dependencies
|
|
|
|
// This ensures we can update the PATH at runtime to get everything loaded
|
|
|
|
|
2024-01-04 16:41:41 +00:00
|
|
|
return newDynamicShimExtServer(AvailableShims["cpu"], model, adapters, projectors, numLayers, opts)
|
2023-12-23 19:35:44 +00:00
|
|
|
}
|