ollama/llm/llama.cpp/patches
2023-10-17 16:55:16 -04:00
..
0001-add-detokenize-endpoint.patch embed libraries using cmake 2023-09-20 14:41:57 -07:00
0001-copy-cuda-runtime-libraries.patch embed libraries using cmake 2023-09-20 14:41:57 -07:00
0001-remove-warm-up-logging.patch Update llama.cpp gguf to latest (#710) 2023-10-17 16:55:16 -04:00
0002-34B-model-support.patch embed libraries using cmake 2023-09-20 14:41:57 -07:00
0003-metal-fix-synchronization-in-new-matrix-multiplicati.patch embed libraries using cmake 2023-09-20 14:41:57 -07:00
0004-metal-add-missing-barriers-for-mul-mat-2699.patch embed libraries using cmake 2023-09-20 14:41:57 -07:00
0005-ggml-support-CUDA-s-half-type-for-aarch64-1455-2670.patch embed libraries using cmake 2023-09-20 14:41:57 -07:00