Fixed .gitignore to ignore any downloaded model .bin files. Cleaned up README.md again
This commit is contained in:
parent
f24e7a7e52
commit
d4eef735d9
2 changed files with 17 additions and 12 deletions
4
.gitignore
vendored
4
.gitignore
vendored
|
@ -165,5 +165,5 @@ cython_debug/
|
||||||
# option (not recommended) you can uncomment the following to ignore the entire idea folder.
|
# option (not recommended) you can uncomment the following to ignore the entire idea folder.
|
||||||
.idea/
|
.idea/
|
||||||
|
|
||||||
# model .bin files
|
# downloaded model .bin files
|
||||||
docker/auto_docker/*.bin
|
docker/open_llama/*.bin
|
||||||
|
|
|
@ -7,15 +7,20 @@
|
||||||
**Note #2:** NVidia GPU CuBLAS support requires a NVidia GPU with sufficient VRAM (approximately as much as the size above) and Docker NVidia support (see [container-toolkit/install-guide](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html))
|
**Note #2:** NVidia GPU CuBLAS support requires a NVidia GPU with sufficient VRAM (approximately as much as the size above) and Docker NVidia support (see [container-toolkit/install-guide](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html))
|
||||||
|
|
||||||
# Simple Dockerfiles for building the llama-cpp-python server with external model bin files
|
# Simple Dockerfiles for building the llama-cpp-python server with external model bin files
|
||||||
- `./openblas_simple/Dockerfile` - a simple Dockerfile for non-GPU OpenBLAS, where the model is located outside the Docker image
|
## openblas_simple - a simple Dockerfile for non-GPU OpenBLAS, where the model is located outside the Docker image
|
||||||
- `cd ./openblas_simple`
|
```
|
||||||
- `docker build -t openblas_simple .`
|
cd ./openblas_simple
|
||||||
- `docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t openblas_simple`
|
docker build -t openblas_simple .
|
||||||
|
docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t openblas_simple
|
||||||
|
```
|
||||||
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
|
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
|
||||||
- `./cuda_simple/Dockerfile` - a simple Dockerfile for CUDA accelerated CuBLAS, where the model is located outside the Docker image
|
|
||||||
- `cd ./cuda_simple`
|
## cuda_simple - a simple Dockerfile for CUDA accelerated CuBLAS, where the model is located outside the Docker image
|
||||||
- `docker build -t cuda_simple .`
|
```
|
||||||
- `docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t cuda_simple`
|
cd ./cuda_simple
|
||||||
|
docker build -t cuda_simple .
|
||||||
|
docker run -e USE_MLOCK=0 -e MODEL=/var/model/<model-path> -v <model-root-path>:/var/model -t cuda_simple
|
||||||
|
```
|
||||||
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
|
where `<model-root-path>/<model-path>` is the full path to the model file on the Docker host system.
|
||||||
|
|
||||||
# "Open-Llama-in-a-box" - Download a MIT licensed Open Llama model and install into a Docker image that runs an OpenBLAS-enabled llama-cpp-python server
|
# "Open-Llama-in-a-box" - Download a MIT licensed Open Llama model and install into a Docker image that runs an OpenBLAS-enabled llama-cpp-python server
|
||||||
|
|
Loading…
Reference in a new issue