# Dockerfiles for building the llama-cpp-python server
-`Dockerfile.openblas_simple` - a simple Dockerfile for non-GPU OpenBLAS
-`Dockerfile.cuda_simple` - a simple Dockerfile for CUDA accelerated CuBLAS
-`hug_model.py` - a Python utility for interactively choosing and downloading the latest `5_1` quantized models from [huggingface.co/TheBloke]( https://huggingface.co/TheBloke)
-`Dockerfile` - a single OpenBLAS and CuBLAS combined Dockerfile that automatically installs a previously downloaded model `model.bin`
Requires a NVidia GPU with sufficient VRAM (approximately as much as the size above) and Docker NVidia support (see [container-toolkit/install-guide](https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/install-guide.html))