ollama/llm/patches
Jeffrey Morgan 152fc202f5
llm: update llama.cpp commit to 7c26775 (#4896)
* llm: update llama.cpp submodule to `7c26775`

* disable `LLAMA_BLAS` for now

* `-DLLAMA_OPENMP=off`
2024-06-17 15:56:16 -04:00
..
01-load-progress.diff llm: update llama.cpp commit to 7c26775 (#4896) 2024-06-17 15:56:16 -04:00
02-clip-log.diff Fix clip log import 2024-04-26 09:43:46 -07:00
03-load_exception.diff bump (#4597) 2024-05-23 14:16:26 -07:00
04-metal.diff use matrix multiplcation kernels in more cases 2024-04-25 13:58:54 -07:00
05-default-pretokenizer.diff llm: update llama.cpp commit to 7c26775 (#4896) 2024-06-17 15:56:16 -04:00
06-qwen2.diff llm: patch to fix qwen 2 temporarily on nvidia (#4897) 2024-06-06 23:14:33 -07:00