From 18ddf6d57d809b22e12f276854b2949f8b52d956 Mon Sep 17 00:00:00 2001 From: Jeffrey Morgan Date: Mon, 8 Jan 2024 20:04:01 -0500 Subject: [PATCH] fix windows build --- llm/ext_server_windows.go | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/llm/ext_server_windows.go b/llm/ext_server_windows.go index 52ee5a6a..39b5f096 100644 --- a/llm/ext_server_windows.go +++ b/llm/ext_server_windows.go @@ -4,9 +4,9 @@ import ( "github.com/jmorganca/ollama/api" ) -func newDefaultExtServer(model string, adapters, projectors []string, numLayers int64, opts api.Options) (extServer, error) { +func newDefaultExtServer(model string, adapters, projectors []string, opts api.Options) (extServer, error) { // On windows we always load the llama.cpp libraries dynamically to avoid startup DLL dependencies // This ensures we can update the PATH at runtime to get everything loaded - return newDynamicShimExtServer(AvailableShims["cpu"], model, adapters, projectors, numLayers, opts) + return newDynamicShimExtServer(AvailableShims["cpu"], model, adapters, projectors, opts) }