Commit graph

  • a0b61ea2a7 Bugfix for models endpoint Andrei Betlen 2023-05-07 20:17:52 -0400
  • e72f58614b Change pointer to lower overhead byref Andrei Betlen 2023-05-07 20:01:34 -0400
  • 14da46f16e Added cache size to settins object. Andrei Betlen 2023-05-07 19:33:17 -0400
  • 0e94a70de1 Add in-memory longest prefix cache. Closes #158 Andrei Betlen 2023-05-07 19:31:26 -0400
  • 8dfde63255 Fix return type Andrei Betlen 2023-05-07 19:30:14 -0400
  • 2753b85321 Format Andrei Betlen 2023-05-07 13:19:56 -0400
  • 4f8cf52a38 Update README Andrei Betlen 2023-05-07 05:20:04 -0400
  • 3adc8fb3ae Update README to use cli options for server Andrei Betlen 2023-05-07 05:10:52 -0400
  • 627811ea83 Add verbose flag to server Andrei Betlen 2023-05-07 05:09:10 -0400
  • 5f43c553d5 Bump version Andrei Betlen 2023-05-07 03:29:33 -0400
  • 3fbda71790 Fix mlock_supported and mmap_supported return type Andrei Betlen 2023-05-07 03:04:22 -0400
  • 5a3413eee3 Update cpu_count Andrei Betlen 2023-05-07 03:03:57 -0400
  • 1a00e452ea Update settings fields and defaults Andrei Betlen 2023-05-07 02:52:20 -0400
  • 86753976c4 Revert "llama_cpp server: delete some ignored / unused parameters" Andrei Betlen 2023-05-07 02:02:34 -0400
  • c382d8f86a Revert "llama_cpp server: mark model as required" Andrei Betlen 2023-05-07 02:00:22 -0400
  • d8fddcce73 Merge branch 'main' of github.com:abetlen/llama_cpp_python into better-server-params-and-fields Andrei Betlen 2023-05-07 01:54:00 -0400
  • 397ae97f64 Update README Andrei Betlen 2023-05-07 01:41:19 -0400
  • 1c188459c4 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-05-07 00:41:10 -0400
  • 7c3743fe5f Update llama.cpp Andrei Betlen 2023-05-07 00:12:47 -0400
  • c76e0913bb
    Update issue templates Andrei 2023-05-06 22:18:31 -0400
  • bc853e3742 Fix type for eval_logits in LlamaState object Andrei Betlen 2023-05-06 21:32:50 -0400
  • ec617c04c1
    Merge pull request #164 from abetlen/dependabot/pip/mkdocstrings-0.21.2 Andrei 2023-05-06 21:30:09 -0400
  • ae3c639764
    Bump mkdocstrings from 0.20.0 to 0.21.2 dependabot[bot] 2023-05-07 00:16:31 +0000
  • 8a692a5838
    Merge pull request #161 from abetlen/dependabot/pip/pytest-7.3.1 Andrei 2023-05-06 20:15:35 -0400
  • 33d41fb8f3
    Bump pytest from 7.2.2 to 7.3.1 dependabot[bot] 2023-05-07 00:07:39 +0000
  • c7becd9363
    Merge pull request #162 from abetlen/dependabot/pip/mkdocs-1.4.3 Andrei 2023-05-06 20:06:14 -0400
  • 515d9bde7e Fixed somethings and activated cublas Maximilian Winter 2023-05-06 23:40:19 +0200
  • 2a21b8f69e
    Bump mkdocs from 1.4.2 to 1.4.3 dependabot[bot] 2023-05-06 21:16:08 +0000
  • d077e641f9
    Merge pull request #165 from abetlen/dependabot/pip/mkdocs-material-9.1.9 Andrei 2023-05-06 17:14:45 -0400
  • fdcab2286c
    Bump mkdocs-material from 9.1.4 to 9.1.9 dependabot[bot] 2023-05-06 21:11:57 +0000
  • bcc4e631cb
    Merge pull request #163 from abetlen/dependabot/pip/black-23.3.0 Andrei 2023-05-06 17:10:30 -0400
  • aa203a0d65 Added mirostat sampling to the high level API. Maximilian Winter 2023-05-06 22:47:47 +0200
  • fd80ddf703 Fix a bug with wrong type Mug 2023-05-06 22:22:28 +0200
  • 996f63e9e1 Add utf8 to chat example Mug 2023-05-06 15:16:58 +0200
  • 3ceb47b597 Fix mirastat requiring c_float Mug 2023-05-06 13:35:50 +0200
  • 9797394c81 Wrong logit_bias parsed type Mug 2023-05-06 13:27:52 +0200
  • 1895c11033 Rename postfix to suffix to match upstream Mug 2023-05-06 13:18:25 +0200
  • c9bb602b26
    Bump black from 23.1.0 to 23.3.0 dependabot[bot] 2023-05-05 23:25:53 +0000
  • 2f2ea00a3d
    Merge pull request #160 from th-neu/main Andrei 2023-05-05 19:24:53 -0400
  • 79d50a29f4
    Create dependabot.yml Thomas Neu 2023-05-06 01:02:59 +0200
  • 980903df93 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-05-05 15:07:26 -0400
  • 98bbd1c6a8 Fix eval logits type Andrei Betlen 2023-05-05 14:23:14 -0400
  • b5f3e74627 Add return type annotations for embeddings and logits Andrei Betlen 2023-05-05 14:22:55 -0400
  • 3e28e0e50c Fix: runtime type errors Andrei Betlen 2023-05-05 14:12:26 -0400
  • e24c3d7447 Prefer explicit imports Andrei Betlen 2023-05-05 14:05:31 -0400
  • 40501435c1 Fix: types Andrei Betlen 2023-05-05 14:04:12 -0400
  • 66e28eb548 Fix temperature bug Andrei Betlen 2023-05-05 14:00:41 -0400
  • 6702d2abfd Fix candidates type Andrei Betlen 2023-05-05 14:00:30 -0400
  • 5e7ddfc3d6 Fix llama_cpp types Andrei Betlen 2023-05-05 13:54:22 -0400
  • f712a04f4e
    Merge pull request #157 from th-neu/th-neu-readme-windows Andrei 2023-05-05 12:40:45 -0400
  • 22c3056b2a
    Update README.md Thomas Neu 2023-05-05 18:40:00 +0200
  • b6a9a0b6ba Add types for all low-level api functions Andrei Betlen 2023-05-05 12:22:27 -0400
  • 5be0efa5f8 Cache should raise KeyError when key is missing Andrei Betlen 2023-05-05 12:21:49 -0400
  • 24fc38754b Add cli options to server. Closes #37 Andrei Betlen 2023-05-05 12:08:28 -0400
  • eb54e30f34
    Update README.md Thomas Neu 2023-05-05 14:22:41 +0200
  • 952ba9ecaf
    Update README.md Thomas Neu 2023-05-05 14:21:57 +0200
  • 5f583b0179 Merge branch 'main' of github.com:abetlen/llama_cpp_python into main Andrei Betlen 2023-05-04 21:59:40 -0400
  • 5c165a85da Bump version Andrei Betlen 2023-05-04 21:59:37 -0400
  • 853dc711cc Format Andrei Betlen 2023-05-04 21:58:36 -0400
  • 97c6372350 Rewind model to longest prefix. Andrei Betlen 2023-05-04 21:58:27 -0400
  • 38b8eeea58
    Merge pull request #154 from th-neu/th-neu-dockerfile-slim Andrei 2023-05-04 19:59:23 -0400
  • 5672ed7fea
    Merge branch 'abetlen:main' into th-neu-dockerfile-slim Thomas Neu 2023-05-04 21:41:13 +0200
  • 501321875f
    Slim-Bullseye based docker image Thomas Neu 2023-05-04 21:03:19 +0200
  • 0e9f227afd Update low level examples Mug 2023-05-04 18:33:08 +0200
  • cabd8b8ed1 Bump version Andrei Betlen 2023-05-04 12:21:20 -0400
  • d78cec67df Update llama.cpp Andrei Betlen 2023-05-04 12:20:25 -0400
  • 329297fafb Bugfix: Missing logits_to_logprobs Andrei Betlen 2023-05-04 12:18:40 -0400
  • d594892fd4 Remove Docker CUDA build job Andrei Betlen 2023-05-04 00:02:46 -0400
  • 0607f6578e Use network installer for cuda Andrei Betlen 2023-05-03 23:22:16 -0400
  • 6d3c20e39d Add CUDA docker image build to github actions Andrei Betlen 2023-05-03 22:20:53 -0400
  • 3008a954c1 Merge branch 'main' of github.com:abetlen/llama-cpp-python into better-server-params-and-fields Lucas Doyle 2023-05-03 13:10:03 -0700
  • a02aa121da Remove cuda build job Andrei Betlen 2023-05-03 10:50:48 -0400
  • 07a56dd9c2 Update job name Andrei Betlen 2023-05-03 10:39:39 -0400
  • 7839eb14d3 Add docker cuda image. Closes #143 Andrei Betlen 2023-05-03 10:29:05 -0400
  • 9e5b6d675a Improve logging messages Andrei Betlen 2023-05-03 10:28:10 -0400
  • 43f2907e3a Support smaller state sizes Andrei Betlen 2023-05-03 09:33:50 -0400
  • 1d47cce222 Update llama.cpp Andrei Betlen 2023-05-03 09:33:30 -0400
  • b9098b0ef7 llama_cpp server: prompt is a string Lucas Doyle 2023-05-02 14:08:51 -0700
  • 0fcc25cdac examples fastapi_server: deprecate Lucas Doyle 2023-04-28 23:54:31 -0700
  • c2e31eecee Update permissions Andrei Betlen 2023-05-02 01:23:17 -0400
  • 63f8d3a6fb Update context Andrei Betlen 2023-05-02 01:16:44 -0400
  • c21a34506e Update permsissions Andrei Betlen 2023-05-02 01:13:43 -0400
  • 872b2ec33f Clone submodules Andrei Betlen 2023-05-02 01:11:34 -0400
  • 62de4692f2 Fix missing dependency Andrei Betlen 2023-05-02 01:09:27 -0400
  • 25062cecd3
    Merge pull request #140 from abetlen/Niek/main Andrei 2023-05-02 01:06:00 -0400
  • 36c81489e7 Remove docker section of publish Andrei Betlen 2023-05-02 01:04:36 -0400
  • 5d5421b29d Add build docker Andrei Betlen 2023-05-02 01:04:02 -0400
  • 81631afc48 Install from local directory Andrei Betlen 2023-05-02 00:55:51 -0400
  • d605408f99 Add dockerignore Andrei Betlen 2023-05-02 00:55:34 -0400
  • e644e75915
    Merge pull request #139 from matthoffner/patch-1 Andrei 2023-05-02 00:33:45 -0400
  • f97ff3c5bb
    Update llama_cpp.py Matt Hoffner 2023-05-01 20:40:06 -0700
  • e9e0654aed Bump version Andrei Betlen 2023-05-01 22:52:25 -0400
  • 7ab08b8d10
    Merge branch 'main' into better-server-params-and-fields Andrei 2023-05-01 22:45:57 -0400
  • 46e3c4b84a Fix Andrei Betlen 2023-05-01 22:41:54 -0400
  • 9eafc4c49a Refactor server to use factory Andrei Betlen 2023-05-01 22:38:46 -0400
  • dd9ad1c759 Formatting Andrei Betlen 2023-05-01 21:51:16 -0400
  • dbbfc4ba2f llama_cpp server: fix to ChatCompletionRequestMessage Lucas Doyle 2023-05-01 11:48:37 -0700
  • fa2a61e065 llama_cpp server: fields for the embedding endpoint Lucas Doyle 2023-04-29 18:46:01 -0700
  • 8dcbf65a45 llama_cpp server: define fields for chat completions Lucas Doyle 2023-04-29 18:37:43 -0700
  • 978b6daf93 llama_cpp server: add some more information to fields for completions Lucas Doyle 2023-04-29 14:37:36 -0700