ollama/llm
Michael Yang 125d0a013a ggufv3
ggufv3 adds support for big endianness, mainly for s390x architecture.
while that's not currently supported for ollama, the change is simple.

loosen version check to be more forward compatible. unless specified,
gguf versions other v1 will be decoded into v2.
2023-10-23 09:35:49 -07:00
..
llama.cpp Update llama.cpp gguf to latest (#710) 2023-10-17 16:55:16 -04:00
falcon.go starcoder 2023-10-02 19:56:51 -07:00
ggml.go ggufv3 2023-10-23 09:35:49 -07:00
gguf.go ggufv3 2023-10-23 09:35:49 -07:00
llama.go simpler check for model loading compatibility errors 2023-10-19 14:50:49 -04:00
llm.go simpler check for model loading compatibility errors 2023-10-19 14:50:49 -04:00
starcoder.go starcoder 2023-10-02 19:56:51 -07:00
utils.go partial decode ggml bin for more info 2023-08-10 09:23:10 -07:00