By default builds will now produce non-debug and non-verbose binaries. To enable verbose logs in llama.cpp and debug symbols in the native code, set `CGO_CFLAGS=-g`
2.1 KiB
Development
- Install cmake or (optionally, required tools for GPUs)
- run
go generate ./...
- run
go build .
Install required tools:
- cmake version 3.24 or higher
- go version 1.20 or higher
- gcc version 11.4.0 or higher
brew install go cmake gcc
Optionally enable debugging and more verbose logging:
export CGO_CFLAGS="-g"
Get the required libraries and build the native LLM code:
go generate ./...
Then build ollama:
go build .
Now you can run ollama
:
./ollama
Building on Linux with GPU support
Linux/Windows CUDA (NVIDIA)
Your operating system distribution may already have packages for NVIDIA CUDA. Distro packages are often preferable, but instructions are distro-specific. Please consult distro-specific docs for dependencies if available!
Install cmake
and golang
as well as NVIDIA CUDA development and runtime packages.
Then generate dependencies:
go generate ./...
Then build the binary:
go build .
Linux ROCm (AMD)
Your operating system distribution may already have packages for AMD ROCm and CLBlast. Distro packages are often preferable, but instructions are distro-specific. Please consult distro-specific docs for dependencies if available!
Install CLBlast and ROCm developement packages first, as well as cmake
and golang
.
Adjust the paths below (correct for Arch) as appropriate for your distributions install locations and generate dependencies:
CLBlast_DIR=/usr/lib/cmake/CLBlast ROCM_PATH=/opt/rocm go generate ./...
Then build the binary:
go build .
ROCm requires elevated privileges to access the GPU at runtime. On most distros you can add your user account to the render
group, or run as root.
Containerized Build
If you have Docker available, you can build linux binaries with ./scripts/build_linux.sh
which has the CUDA and ROCm dependencies included.