Jeffrey Morgan
c416087339
import.md
: formatting and spelling
2023-10-15 01:39:46 -04:00
Jeffrey Morgan
6002cebd2c
import.md
: convert and quantize docs
2023-10-15 00:11:51 -04:00
Jeffrey Morgan
212bdc541c
import.md
: model architectures spelling
2023-10-15 00:07:58 -04:00
Jeffrey Morgan
dca6686273
add steps for creating a Modelfile and more example commands to import.md
2023-10-15 00:05:50 -04:00
Matt Williams
b2974a7095
applied mikes comments
...
Signed-off-by: Matt Williams <m@technovangelist.com>
2023-10-14 08:29:24 -07:00
Matt Williams
3c975f898f
update doc to refer to docker image
...
Signed-off-by: Matt Williams <m@technovangelist.com>
2023-10-12 15:57:50 -07:00
Matt Williams
9245c8a1df
add how to quantize doc
...
Signed-off-by: Matt Williams <m@technovangelist.com>
2023-10-12 15:34:57 -07:00
Bruce MacDonald
274d5a5fdf
optional parameter to not stream response ( #639 )
...
* update streaming request accept header
* add optional stream param to request bodies
2023-10-11 12:54:27 -04:00
Costa Alexoglou
f7f5169c94
Update api.md ( #741 )
...
Avoid triple ticks in visual editor and also copied in clipboard.
2023-10-09 16:01:46 -04:00
James Braza
6f2ce74231
Got rif of all caps to show it can be lower case
2023-10-02 13:54:27 -07:00
James Braza
6edcc5c79f
Using code highlighting syntax around Modelfile
2023-10-02 13:46:05 -07:00
Jiayu Liu
4fc10acce9
add some missing code directives in docs ( #664 )
2023-10-01 11:51:01 -07:00
Jay Nakrani
1d0ebe67e8
Document response stream chunk delimiter. ( #632 )
...
Document response stream chunk delimiter.
2023-09-29 21:45:52 -07:00
Aaron Coffey
6ae33d8141
Update modelfile.md to reflect the usage of num_gpu. ( #629 )
2023-09-28 10:21:21 -04:00
Jeffrey Morgan
c5664c1fef
Update faq.md
2023-09-27 13:49:43 -07:00
Bruce MacDonald
ed20837f9a
Update modelfile.md
2023-09-27 10:38:10 -04:00
James Braza
1db2a61dd0
Added num_predict to the options table ( #614 )
2023-09-27 10:26:08 -04:00
Jeffrey Morgan
5306b0269d
Update linux.md
2023-09-25 16:10:32 -07:00
Jeffrey Morgan
0fb5268496
Update linux.md
2023-09-25 10:06:23 -07:00
Jeffrey Morgan
ee3032ad89
improvements to docs/linux.md
2023-09-24 21:50:07 -07:00
Jeffrey Morgan
5b7a27281d
improvements to docs/linux.md
2023-09-24 21:38:23 -07:00
Jeffrey Morgan
d2a784e33e
add docs/linux.md
2023-09-24 21:34:44 -07:00
Michael Yang
6c6a31a1e8
embed libraries using cmake
2023-09-20 14:41:57 -07:00
Bruce MacDonald
fc6ec356fc
remove libcuda.so
2023-09-20 20:36:14 +01:00
Bruce MacDonald
1255bc9b45
only package 11.8 runner
2023-09-20 20:00:41 +01:00
Bruce MacDonald
4e8be787c7
pack in cuda libs
2023-09-20 17:40:42 +01:00
Bruce MacDonald
2540c9181c
support for packaging in multiple cuda runners ( #509 )
...
* enable packaging multiple cuda versions
* use nvcc cuda version if available
---------
Co-authored-by: Michael Yang <mxyng@pm.me>
2023-09-14 15:08:13 -04:00
Matt Williams
fc8707686f
Update API docs ( #527 )
...
* Update API docs
Signed-off-by: Matt Williams <m@technovangelist.com>
* strange TOC was getting auto generated
Signed-off-by: Matt Williams <m@technovangelist.com>
* Update docs/api.md
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
* Update docs/api.md
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
* Update docs/api.md
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
* Update api.md
---------
Signed-off-by: Matt Williams <m@technovangelist.com>
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
Co-authored-by: Michael Chiang <mchiang0610@users.noreply.github.com>
2023-09-14 08:51:26 -07:00
Bruce MacDonald
f221637053
first pass at linux gpu support ( #454 )
...
* linux gpu support
* handle multiple gpus
* add cuda docker image (#488 )
---------
Co-authored-by: Michael Yang <mxyng@pm.me>
2023-09-12 11:04:35 -04:00
Ackermann Yuriy
154f24af91
Added missing options params to the embeddings docs ( #472 )
2023-09-05 20:18:49 -04:00
Bruce MacDonald
42998d797d
subprocess llama.cpp server ( #401 )
...
* remove c code
* pack llama.cpp
* use request context for llama_cpp
* let llama_cpp decide the number of threads to use
* stop llama runner when app stops
* remove sample count and duration metrics
* use go generate to get libraries
* tmp dir for running llm
2023-08-30 16:35:03 -04:00
Quinn Slack
f4432e1dba
treat stop as stop sequences, not exact tokens ( #442 )
...
The `stop` option to the generate API is a list of sequences that should cause generation to stop. Although these are commonly called "stop tokens", they do not necessarily correspond to LLM tokens (per the LLM's tokenizer). For example, if the caller sends a generate request with `"stop":["\n"]`, then generation should stop on any token containing `\n` (and trim `\n` from the output), not just if the token exactly matches `\n`. If `stop` were interpreted strictly as LLM tokens, then it would require callers of the generate API to know the LLM's tokenizer and enumerate many tokens in the `stop` list.
Fixes https://github.com/jmorganca/ollama/issues/295 .
2023-08-30 11:53:42 -04:00
Jeffrey Morgan
d3b838ce60
update orca
to orca-mini
2023-08-27 13:26:30 -04:00
Michael Yang
041f9ad1a1
update README.md
2023-08-25 11:44:25 -07:00
Bruce MacDonald
519f4d98ef
add embed docs for modelfile
2023-08-17 13:37:42 -04:00
Bruce MacDonald
23e1da778d
Add context to api docs
2023-08-15 11:43:22 -03:00
Bruce MacDonald
53bc36d207
Update modelfile.md
2023-08-15 09:23:36 -03:00
Bruce MacDonald
af98a1773f
update python example
2023-08-14 16:38:44 -03:00
Bruce MacDonald
9ae9a89883
Update modelfile.md
2023-08-14 16:26:53 -03:00
Bruce MacDonald
648f0974c6
python example
2023-08-14 15:27:13 -03:00
Bruce MacDonald
fc5230dffa
Add context to api docs
2023-08-14 15:23:24 -03:00
Güvenç Usanmaz
4c33a9ac67
Update langchainpy.md
...
base_url value for Ollama object creation is corrected.
2023-08-14 12:12:56 +03:00
Matt Williams
202c29c21a
resolving bmacd comment
...
Signed-off-by: Matt Williams <m@technovangelist.com>
2023-08-11 13:51:44 -07:00
Matt Williams
c1c871620a
Update docs/tutorials/langchainjs.md
...
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
2023-08-11 13:48:46 -07:00
Matt Williams
a21a8bef56
Update docs/tutorials/langchainjs.md
...
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
2023-08-11 13:48:35 -07:00
Matt Williams
522726228a
Update docs/tutorials.md
...
Co-authored-by: Bruce MacDonald <brucewmacdonald@gmail.com>
2023-08-11 13:48:16 -07:00
Matt Williams
d3ee1329e9
Add tutorials for using Langchain with ollama
...
Signed-off-by: Matt Williams <m@technovangelist.com>
2023-08-10 21:27:37 -07:00
Michael Yang
3a05d3def7
Merge pull request #326 from asarturas/document-num-gqa-parameter
...
Document num_gqa parameter
2023-08-10 18:18:38 -07:00
Arturas Smorgun
d9c2687fd0
document default num_gqa to 1, as it's applicable to most models
...
Co-authored-by: Michael Yang <mxyng@pm.me>
2023-08-11 01:29:40 +01:00
Michael Yang
6517bcc53c
Merge pull request #290 from jmorganca/add-adapter-layers
...
implement loading ggml lora adapters through the modelfile
2023-08-10 17:23:01 -07:00