1.9 KiB
FAQ
How can I view the logs?
On macOS:
cat ~/.ollama/logs/server.log
On Linux:
journalctl -u ollama
If you're running ollama serve
directly, the logs will be printed to the console.
How can I expose Ollama on my network?
Ollama binds to 127.0.0.1 port 11434 by default. Change the bind address with the OLLAMA_HOST
environment variable.
On macOS:
OLLAMA_HOST=0.0.0.0:11435 ollama serve
On Linux:
Create a systemd
drop-in directory and set Environment=OLLAMA_HOST
mkdir -p /etc/systemd/system/ollama.service.d
echo "[Service]" >>/etc/systemd/system/ollama.service.d/environment.conf
echo "Environment=OLLAMA_HOST=0.0.0.0:11434" >>/etc/systemd/system/ollama.service.d/environment.conf
Reload systemd
and restart Ollama:
systemctl daemon-reload
systemctl restart ollama
How can I allow additional web origins to access Ollama?
Ollama allows cross origin requests from 127.0.0.1
and 0.0.0.0
by default. Add additional origins with the OLLAMA_ORIGINS
environment variable:
On macOS:
OLLAMA_ORIGINS=http://192.168.1.1:*,https://example.com ollama serve
On Linux:
echo "Environment=OLLAMA_ORIGINS=http://129.168.1.1:*,https://example.com" >>/etc/systemd/system/ollama.service.d/environment.conf
Reload systemd
and restart Ollama:
systemctl daemon-reload
systemctl restart ollama
Where are models stored?
- macOS: Raw model data is stored under
~/.ollama/models
. - Linux: Raw model data is stored under
/usr/share/ollama/.ollama/models
How can I change where Ollama stores models?
To modify where models are stored, you can use the OLLAMA_MODELS
environment variable. Note that on Linux this means defining OLLAMA_MODELS
in a drop-in /etc/systemd/system/ollama.service.d
service file, reloading systemd, and restarting the ollama service.