Use llama2
as the model in api.md
This commit is contained in:
parent
41434a7cdc
commit
92656a74b7
1 changed files with 9 additions and 9 deletions
18
docs/api.md
18
docs/api.md
|
@ -114,7 +114,7 @@ To calculate how fast the response is generated in tokens per second (token/s),
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
curl -X POST http://localhost:11434/api/generate -d '{
|
curl -X POST http://localhost:11434/api/generate -d '{
|
||||||
"model": "llama2:7b",
|
"model": "llama2",
|
||||||
"prompt": "Why is the sky blue?",
|
"prompt": "Why is the sky blue?",
|
||||||
"stream": false
|
"stream": false
|
||||||
}'
|
}'
|
||||||
|
@ -126,7 +126,7 @@ If `stream` is set to `false`, the response will be a single JSON object:
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"model": "llama2:7b",
|
"model": "llama2",
|
||||||
"created_at": "2023-08-04T19:22:45.499127Z",
|
"created_at": "2023-08-04T19:22:45.499127Z",
|
||||||
"response": "The sky is blue because it is the color of the sky.",
|
"response": "The sky is blue because it is the color of the sky.",
|
||||||
"context": [1, 2, 3],
|
"context": [1, 2, 3],
|
||||||
|
@ -225,7 +225,7 @@ If you want to set custom options for the model at runtime rather than in the Mo
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
curl -X POST http://localhost:11434/api/generate -d '{
|
curl -X POST http://localhost:11434/api/generate -d '{
|
||||||
"model": "llama2:7b",
|
"model": "llama2",
|
||||||
"prompt": "Why is the sky blue?",
|
"prompt": "Why is the sky blue?",
|
||||||
"stream": false,
|
"stream": false,
|
||||||
"options": {
|
"options": {
|
||||||
|
@ -270,7 +270,7 @@ curl -X POST http://localhost:11434/api/generate -d '{
|
||||||
|
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"model": "llama2:7b",
|
"model": "llama2",
|
||||||
"created_at": "2023-08-04T19:22:45.499127Z",
|
"created_at": "2023-08-04T19:22:45.499127Z",
|
||||||
"response": "The sky is blue because it is the color of the sky.",
|
"response": "The sky is blue because it is the color of the sky.",
|
||||||
"context": [1, 2, 3],
|
"context": [1, 2, 3],
|
||||||
|
@ -395,7 +395,7 @@ A single JSON object will be returned.
|
||||||
{
|
{
|
||||||
"models": [
|
"models": [
|
||||||
{
|
{
|
||||||
"name": "llama2:7b",
|
"name": "llama2",
|
||||||
"modified_at": "2023-08-02T17:02:23.713454393-07:00",
|
"modified_at": "2023-08-02T17:02:23.713454393-07:00",
|
||||||
"size": 3791730596
|
"size": 3791730596
|
||||||
},
|
},
|
||||||
|
@ -426,7 +426,7 @@ Show details about a model including modelfile, template, parameters, license, a
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
curl http://localhost:11434/api/show -d '{
|
curl http://localhost:11434/api/show -d '{
|
||||||
"name": "llama2:7b"
|
"name": "llama2"
|
||||||
}'
|
}'
|
||||||
```
|
```
|
||||||
|
|
||||||
|
@ -455,7 +455,7 @@ Copy a model. Creates a model with another name from an existing model.
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
curl http://localhost:11434/api/copy -d '{
|
curl http://localhost:11434/api/copy -d '{
|
||||||
"source": "llama2:7b",
|
"source": "llama2",
|
||||||
"destination": "llama2-backup"
|
"destination": "llama2-backup"
|
||||||
}'
|
}'
|
||||||
```
|
```
|
||||||
|
@ -510,7 +510,7 @@ Download a model from the ollama library. Cancelled pulls are resumed from where
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
curl -X POST http://localhost:11434/api/pull -d '{
|
curl -X POST http://localhost:11434/api/pull -d '{
|
||||||
"name": "llama2:7b"
|
"name": "llama2"
|
||||||
}'
|
}'
|
||||||
```
|
```
|
||||||
|
|
||||||
|
@ -650,7 +650,7 @@ Advanced parameters:
|
||||||
|
|
||||||
```shell
|
```shell
|
||||||
curl -X POST http://localhost:11434/api/embeddings -d '{
|
curl -X POST http://localhost:11434/api/embeddings -d '{
|
||||||
"model": "llama2:7b",
|
"model": "llama2",
|
||||||
"prompt": "Here is an article about llamas..."
|
"prompt": "Here is an article about llamas..."
|
||||||
}'
|
}'
|
||||||
```
|
```
|
||||||
|
|
Loading…
Reference in a new issue