set(TARGET ext_server) option(LLAMA_SERVER_VERBOSE "Build verbose logging option for Server" ON) if (WIN32) add_library(${TARGET} SHARED ext_server.cpp ../llama.cpp/llama.cpp) else() add_library(${TARGET} STATIC ext_server.cpp ../llama.cpp/llama.cpp) endif() target_compile_features(${TARGET} PRIVATE cxx_std_11) target_compile_definitions(${TARGET} PUBLIC LLAMA_SERVER_LIBRARY=1) target_link_libraries(${TARGET} PRIVATE ggml llava common ) set_target_properties(${TARGET} PROPERTIES POSITION_INDEPENDENT_CODE ON) target_compile_definitions(${TARGET} PRIVATE SERVER_VERBOSE=$) install(TARGETS ext_server LIBRARY) if (CUDAToolkit_FOUND) target_include_directories(${TARGET} PRIVATE ${CMAKE_CUDA_TOOLKIT_INCLUDE_DIRECTORIES}) if (WIN32) target_link_libraries(${TARGET} PRIVATE nvml) endif() endif()