ollama/llm/ext_server
Daniel Hiltgen 90ca84172c
Fix embeddings memory corruption (#6467)
* Fix embeddings memory corruption

The patch was leading to a buffer overrun corruption.  Once removed though, parallism
in server.cpp lead to hitting an assert due to slot/seq IDs being >= token count.  To
work around this, only use slot 0 for embeddings.

* Fix embed integration test assumption

The token eval count has changed with recent llama.cpp bumps (0.3.5+)
2024-08-22 14:51:42 -07:00
..
CMakeLists.txt Refactor linux packaging 2024-08-19 09:38:53 -07:00
httplib.h Import server.cpp as of b2356 2024-03-12 13:58:06 -07:00
json.hpp Import server.cpp as of b2356 2024-03-12 13:58:06 -07:00
server.cpp Fix embeddings memory corruption (#6467) 2024-08-22 14:51:42 -07:00
utils.hpp log clean up 2024-05-09 14:55:36 -07:00