llama.cpp/docker/README.md

58 lines
3 KiB
Markdown
Raw Normal View History

# Simple Dockerfiles for building the llama-cpp-python server with external model bin files
- `./openblas_simple/Dockerfile` - a simple Dockerfile for non-GPU OpenBLAS, where the model is located outside the Docker image
- `cd ./openblas_simple`
- `docker build -t openblas_simple .`
- `docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t openblas_simple`
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
- `./cuda_simple/Dockerfile` - a simple Dockerfile for CUDA accelerated CuBLAS, where the model is located outside the Docker image
- `cd ./cuda_simple`
- `docker build -t cuda_simple .`
- `docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t cuda_simple`
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
# "Bot-in-a-box" - a method to build a Docker image by choosing a model to be downloaded and loading into a Docker image
- `cd ./auto_docker`:
- `hug_model.py` - a Python utility for interactively choosing and downloading the latest `5_1` quantized models from [huggingface.co/TheBloke]( https://huggingface.co/TheBloke)
2023-05-23 19:50:39 +00:00
- `Dockerfile` - a single OpenBLAS and CuBLAS combined Dockerfile that automatically installs a previously downloaded model `model.bin`
## Get model from Hugging Face
2023-05-23 19:26:40 +00:00
`python3 ./hug_model.py`
2023-05-23 19:36:21 +00:00
You should now have a model in the current directory and `model.bin` symlinked to it for the subsequent Docker build and copy step. e.g.
2023-05-23 19:26:40 +00:00
```
docker $ ls -lh *.bin
2023-05-23 19:36:21 +00:00
-rw-rw-r-- 1 user user 4.8G May 23 18:30 <downloaded-model-file>.q5_1.bin
2023-05-23 19:26:40 +00:00
lrwxrwxrwx 1 user user 24 May 23 18:30 model.bin -> <downloaded-model-file>.q5_1.bin
```
2023-05-23 19:50:39 +00:00
**Note #1:** Make sure you have enough disk space to download the model. As the model is then copied into the image you will need at least
2023-05-23 19:26:40 +00:00
**TWICE** as much disk space as the size of the model:
| Model | Quantized size |
|------:|----------------:|
| 7B | 5 GB |
| 13B | 10 GB |
| 30B | 25 GB |
| 65B | 50 GB |
2023-05-23 19:36:21 +00:00
**Note #2:** If you want to pass or tune additional parameters, customise `./start_server.sh` before running `docker build ...`
2023-05-23 19:26:40 +00:00
2023-05-23 19:36:21 +00:00
# Install Docker Server
2023-05-23 19:50:39 +00:00
**Note #3:** This was tested with Docker running on Linux. If you can get it working on Windows or MacOS, please update this `README.md` with a PR!
2023-05-23 19:36:21 +00:00
[Install Docker Engine](https://docs.docker.com/engine/install)
# Use OpenBLAS
2023-05-23 19:50:39 +00:00
Use if you don't have a NVidia GPU. Defaults to `python:3-slim-bullseye` Docker base image and OpenBLAS:
2023-05-23 19:26:40 +00:00
## Build:
`docker build --build-arg -t openblas .`
## Run:
`docker run --cap-add SYS_RESOURCE -t openblas`
# Use CuBLAS
2023-05-23 19:50:39 +00:00
Requires a NVidia GPU with sufficient VRAM (approximately as much as the size above) and Docker NVidia support (see [container-toolkit/install-guide](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html))
2023-05-23 19:26:40 +00:00
## Build:
2023-05-23 19:50:39 +00:00
`docker build --build-arg IMAGE=nvidia/cuda:12.1.1-devel-ubuntu22.04 -t cublas .`
2023-05-23 19:26:40 +00:00
## Run:
`docker run --cap-add SYS_RESOURCE -t cublas`