Signed-off-by: Matt Williams <m@technovangelist.com>
2.7 KiB
FAQ
How can I view the logs?
On macOS:
cat ~/.ollama/logs/server.log
On Linux:
journalctl -u ollama
If you're running ollama serve
directly, the logs will be printed to the console.
How can I expose Ollama on my network?
Ollama binds to 127.0.0.1 port 11434 by default. Change the bind address with the OLLAMA_HOST
environment variable.
On macOS:
OLLAMA_HOST=0.0.0.0:11435 ollama serve
On Linux:
Create a systemd
drop-in directory and set Environment=OLLAMA_HOST
mkdir -p /etc/systemd/system/ollama.service.d
echo "[Service]" >>/etc/systemd/system/ollama.service.d/environment.conf
echo "Environment=OLLAMA_HOST=0.0.0.0:11434" >>/etc/systemd/system/ollama.service.d/environment.conf
Reload systemd
and restart Ollama:
systemctl daemon-reload
systemctl restart ollama
How can I allow additional web origins to access Ollama?
Ollama allows cross origin requests from 127.0.0.1
and 0.0.0.0
by default. Add additional origins with the OLLAMA_ORIGINS
environment variable:
On macOS:
OLLAMA_ORIGINS=http://192.168.1.1:*,https://example.com ollama serve
On Linux:
echo "Environment=OLLAMA_ORIGINS=http://129.168.1.1:*,https://example.com" >>/etc/systemd/system/ollama.service.d/environment.conf
Reload systemd
and restart Ollama:
systemctl daemon-reload
systemctl restart ollama
Where are models stored?
- macOS: Raw model data is stored under
~/.ollama/models
. - Linux: Raw model data is stored under
/usr/share/ollama/.ollama/models
Below the models directory you will find a structure similar to the following:
.
├── blobs
└── manifests
└── registry.ollama.ai
├── f0rodo
├── library
├── mattw
└── saikatkumardey
There is a manifests/registry.ollama.ai/namespace
path. In example above, the user has downloaded models from the official library
, f0rodo
, mattw
, and saikatkumardey
namespaces. Within each of those directories, you will find directories for each of the models downloaded. And in there you will find a file name representing each tag. Each tag file is the manifest for the model.
The manifest lists all the layers used in this model. You will see a media type
for each layer, along with a digest. That digest corresponds with a file in the models/blobs directory
.
How can I change where Ollama stores models?
To modify where models are stored, you can use the OLLAMA_MODELS
environment variable. Note that on Linux this means defining OLLAMA_MODELS
in a drop-in /etc/systemd/system/ollama.service.d
service file, reloading systemd, and restarting the ollama service.