1.5 KiB
1.5 KiB
Installing Ollama on Linux
Note: A one line installer for Ollama is available by running:
curl https://ollama.ai/install.sh | sh
Download the ollama
binary
Ollama is distributed as a self-contained binary. Download it to a directory in your PATH:
sudo curl -L https://ollama.ai/download/ollama-linux-amd64 -o /usr/bin/ollama
sudo chmod +x /usr/bin/ollama
Start Ollama
Start Ollama by running ollama serve
:
ollama serve
Once Ollama is running, run a model in another terminal session:
ollama run llama2
Install CUDA drivers (optional – for Nvidia GPUs)
Download and install CUDA.
Verify that the drivers are installed by running the following command, which should print details about your GPU:
nvidia-smi
Adding Ollama as a startup service (optional)
Create a user for Ollama:
sudo useradd -r -s /bin/false -m -d /usr/share/ollama ollama
Create a service file in /etc/systemd/system/ollama.service
:
[Unit]
Description=Ollama Service
After=network-online.target
[Service]
ExecStart=/usr/bin/ollama serve
User=ollama
Group=ollama
Restart=always
RestartSec=3
Environment="HOME=/usr/share/ollama"
[Install]
WantedBy=default.target
Then start the service:
sudo systemctl daemon-reload
sudo systemctl enable ollama
Viewing logs
To view logs of Ollama running as a startup service, run:
journalctl -u ollama