ollama/docs/linux.md
2023-09-24 21:50:07 -07:00

1.4 KiB
Raw Blame History

Installing Ollama on Linux

Note: A one line installer for Ollama is available by running:

curl https://ollama.ai/install.sh | sh

Download the ollama binary

Ollama is distributed as a self-contained binary. Download it to a directory in your PATH:

sudo curl -L https://ollama.ai/download/ollama-linux-amd64 -o /usr/bin/ollama
sudo chmod +x /usr/bin/ollama

Start Ollama

Start Ollama by running ollama serve:

ollama serve

Once Ollama is running, run a model in another terminal session:

ollama run llama2

Install CUDA drivers (optional for Nvidia GPUs)

Download and install CUDA.

Verify that the drivers are installed by running the following command, which should print details about your GPU:

nvidia-smi

Adding Ollama as a startup service (optional)

Create a user for Ollama:

sudo useradd -r -s /bin/false -m -d /usr/share/ollama ollama

Create a service file in /etc/systemd/system/ollama.service:

[Unit]
Description=Ollama Service
After=network-online.target

[Service]
ExecStart=/usr/bin/ollama serve
User=ollama
Group=ollama
Restart=always
RestartSec=3
Environment="HOME=/usr/share/ollama"

[Install]
WantedBy=default.target

Then start the service:

sudo systemctl daemon-reload
sudo systemctl enable ollama

Run a model

ollama run llama2