Merge branch 'main' of https://github.com/abetlen/llama-cpp-python into main
This commit is contained in:
commit
6981597835
2 changed files with 2 additions and 1 deletions
|
@ -14,6 +14,7 @@ This package provides:
|
||||||
- High-level Python API for text completion
|
- High-level Python API for text completion
|
||||||
- OpenAI-like API
|
- OpenAI-like API
|
||||||
- [LangChain compatibility](https://python.langchain.com/docs/integrations/llms/llamacpp)
|
- [LangChain compatibility](https://python.langchain.com/docs/integrations/llms/llamacpp)
|
||||||
|
- [LlamaIndex compatibility](https://docs.llamaindex.ai/en/stable/examples/llm/llama_2_llama_cpp.html)
|
||||||
- OpenAI compatible web server
|
- OpenAI compatible web server
|
||||||
- [Local Copilot replacement](https://llama-cpp-python.readthedocs.io/en/latest/server/#code-completion)
|
- [Local Copilot replacement](https://llama-cpp-python.readthedocs.io/en/latest/server/#code-completion)
|
||||||
- [Function Calling support](https://llama-cpp-python.readthedocs.io/en/latest/server/#function-calling)
|
- [Function Calling support](https://llama-cpp-python.readthedocs.io/en/latest/server/#function-calling)
|
||||||
|
|
|
@ -60,7 +60,7 @@ class ModelSettings(BaseSettings):
|
||||||
seed: int = Field(
|
seed: int = Field(
|
||||||
default=llama_cpp.LLAMA_DEFAULT_SEED, description="Random seed. -1 for random."
|
default=llama_cpp.LLAMA_DEFAULT_SEED, description="Random seed. -1 for random."
|
||||||
)
|
)
|
||||||
n_ctx: int = Field(default=2048, ge=1, description="The context size.")
|
n_ctx: int = Field(default=2048, ge=0, description="The context size.")
|
||||||
n_batch: int = Field(
|
n_batch: int = Field(
|
||||||
default=512, ge=1, description="The batch size to use per eval."
|
default=512, ge=1, description="The batch size to use per eval."
|
||||||
)
|
)
|
||||||
|
|
Loading…
Reference in a new issue