api: start adding documentation to package api (#2878)
* api: start adding documentation to package api Updates #2840 * Fix lint typo report
This commit is contained in:
parent
4340f8eba4
commit
ad90b9ab3d
2 changed files with 74 additions and 10 deletions
|
@ -1,3 +1,9 @@
|
||||||
|
// Package api implements the client-side API for code wishing to interact
|
||||||
|
// with the ollama service. The methods of the [Client] type correspond to
|
||||||
|
// the ollama REST API as described in https://github.com/ollama/ollama/blob/main/docs/api.md
|
||||||
|
//
|
||||||
|
// The ollama command-line client itself uses this package to interact with
|
||||||
|
// the backend service.
|
||||||
package api
|
package api
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
@ -18,6 +24,8 @@ import (
|
||||||
"github.com/ollama/ollama/version"
|
"github.com/ollama/ollama/version"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Client encapsulates client state for interacting with the ollama
|
||||||
|
// service. Use [ClientFromEnvironment] to create new Clients.
|
||||||
type Client struct {
|
type Client struct {
|
||||||
base *url.URL
|
base *url.URL
|
||||||
http *http.Client
|
http *http.Client
|
||||||
|
@ -39,6 +47,15 @@ func checkError(resp *http.Response, body []byte) error {
|
||||||
return apiError
|
return apiError
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ClientFromEnvironment creates a new [Client] using configuration from the
|
||||||
|
// environment variable OLLAMA_HOST, which points to the network host and
|
||||||
|
// port on which the ollama service is listenting. The format of this variable
|
||||||
|
// is:
|
||||||
|
//
|
||||||
|
// <scheme>://<host>:<port>
|
||||||
|
//
|
||||||
|
// If the variable is not specified, a default ollama host and port will be
|
||||||
|
// used.
|
||||||
func ClientFromEnvironment() (*Client, error) {
|
func ClientFromEnvironment() (*Client, error) {
|
||||||
defaultPort := "11434"
|
defaultPort := "11434"
|
||||||
|
|
||||||
|
@ -190,8 +207,14 @@ func (c *Client) stream(ctx context.Context, method, path string, data any, fn f
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GenerateResponseFunc is a function that [Client.Generate] invokes every time
|
||||||
|
// a response is received from the service. If this function returns an error,
|
||||||
|
// [Client.Generate] will stop generating and return this error.
|
||||||
type GenerateResponseFunc func(GenerateResponse) error
|
type GenerateResponseFunc func(GenerateResponse) error
|
||||||
|
|
||||||
|
// Generate generates a response for a given prompt. The req parameter should
|
||||||
|
// be populated with prompt details. fn is called for each response (there may
|
||||||
|
// be multiple responses, e.g. in case streaming is enabled).
|
||||||
func (c *Client) Generate(ctx context.Context, req *GenerateRequest, fn GenerateResponseFunc) error {
|
func (c *Client) Generate(ctx context.Context, req *GenerateRequest, fn GenerateResponseFunc) error {
|
||||||
return c.stream(ctx, http.MethodPost, "/api/generate", req, func(bts []byte) error {
|
return c.stream(ctx, http.MethodPost, "/api/generate", req, func(bts []byte) error {
|
||||||
var resp GenerateResponse
|
var resp GenerateResponse
|
||||||
|
@ -203,8 +226,15 @@ func (c *Client) Generate(ctx context.Context, req *GenerateRequest, fn Generate
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ChatResponseFunc is a function that [Client.Chat] invokes every time
|
||||||
|
// a response is received from the service. If this function returns an error,
|
||||||
|
// [Client.Chat] will stop generating and return this error.
|
||||||
type ChatResponseFunc func(ChatResponse) error
|
type ChatResponseFunc func(ChatResponse) error
|
||||||
|
|
||||||
|
// Chat generates the next message in a chat. [ChatRequest] may contain a
|
||||||
|
// sequence of messages which can be used to maintain chat history with a model.
|
||||||
|
// fn is called for each response (there may be multiple responses, e.g. if case
|
||||||
|
// streaming is enabled).
|
||||||
func (c *Client) Chat(ctx context.Context, req *ChatRequest, fn ChatResponseFunc) error {
|
func (c *Client) Chat(ctx context.Context, req *ChatRequest, fn ChatResponseFunc) error {
|
||||||
return c.stream(ctx, http.MethodPost, "/api/chat", req, func(bts []byte) error {
|
return c.stream(ctx, http.MethodPost, "/api/chat", req, func(bts []byte) error {
|
||||||
var resp ChatResponse
|
var resp ChatResponse
|
||||||
|
@ -216,8 +246,14 @@ func (c *Client) Chat(ctx context.Context, req *ChatRequest, fn ChatResponseFunc
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// PullProgressFunc is a function that [Client.Pull] invokes every time there
|
||||||
|
// is progress with a "pull" request sent to the service. If this function
|
||||||
|
// returns an error, [Client.Pull] will stop the process and return this error.
|
||||||
type PullProgressFunc func(ProgressResponse) error
|
type PullProgressFunc func(ProgressResponse) error
|
||||||
|
|
||||||
|
// Pull downloads a model from the ollama library. fn is called each time
|
||||||
|
// progress is made on the request and can be used to display a progress bar,
|
||||||
|
// etc.
|
||||||
func (c *Client) Pull(ctx context.Context, req *PullRequest, fn PullProgressFunc) error {
|
func (c *Client) Pull(ctx context.Context, req *PullRequest, fn PullProgressFunc) error {
|
||||||
return c.stream(ctx, http.MethodPost, "/api/pull", req, func(bts []byte) error {
|
return c.stream(ctx, http.MethodPost, "/api/pull", req, func(bts []byte) error {
|
||||||
var resp ProgressResponse
|
var resp ProgressResponse
|
||||||
|
|
28
api/types.go
28
api/types.go
|
@ -33,18 +33,46 @@ func (e StatusError) Error() string {
|
||||||
|
|
||||||
type ImageData []byte
|
type ImageData []byte
|
||||||
|
|
||||||
|
// GenerateRequest describes a request sent by [Client.Generate]. While you
|
||||||
|
// have to specify the Model and Prompt fields, all the other fields have
|
||||||
|
// reasonable defaults for basic uses.
|
||||||
type GenerateRequest struct {
|
type GenerateRequest struct {
|
||||||
|
// Model is the model name; it should be a name familiar to Ollama from
|
||||||
|
// the library at https://ollama.com/library
|
||||||
Model string `json:"model"`
|
Model string `json:"model"`
|
||||||
|
|
||||||
|
// Prompt is the textual prompt to send to the model.
|
||||||
Prompt string `json:"prompt"`
|
Prompt string `json:"prompt"`
|
||||||
|
|
||||||
|
// System overrides the model's default system message/prompt.
|
||||||
System string `json:"system"`
|
System string `json:"system"`
|
||||||
|
|
||||||
|
// Template overrides the model's default prompt template.
|
||||||
Template string `json:"template"`
|
Template string `json:"template"`
|
||||||
|
|
||||||
|
// Context is the context parameter returned from a previous call to
|
||||||
|
// Generate call. It can be used to keep a short conversational memory.
|
||||||
Context []int `json:"context,omitempty"`
|
Context []int `json:"context,omitempty"`
|
||||||
|
|
||||||
|
// Stream specifies whether the response is streaming; it is true by default.
|
||||||
Stream *bool `json:"stream,omitempty"`
|
Stream *bool `json:"stream,omitempty"`
|
||||||
|
|
||||||
|
// Raw set to true means that no formatting will be applied to the prompt.
|
||||||
Raw bool `json:"raw,omitempty"`
|
Raw bool `json:"raw,omitempty"`
|
||||||
|
|
||||||
|
// Format specifies the format to return a response in.
|
||||||
Format string `json:"format"`
|
Format string `json:"format"`
|
||||||
|
|
||||||
|
// KeepAlive controls how long the model will stay loaded in memory following
|
||||||
|
// this request.
|
||||||
KeepAlive *Duration `json:"keep_alive,omitempty"`
|
KeepAlive *Duration `json:"keep_alive,omitempty"`
|
||||||
|
|
||||||
|
// Images is an optional list of base64-encoded images accompanying this
|
||||||
|
// request, for multimodal models.
|
||||||
Images []ImageData `json:"images,omitempty"`
|
Images []ImageData `json:"images,omitempty"`
|
||||||
|
|
||||||
|
// Options lists model-specific options. For example, temperature can be
|
||||||
|
// set through this field, if the model supports it.
|
||||||
Options map[string]interface{} `json:"options"`
|
Options map[string]interface{} `json:"options"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue