From 408dd14e5b0a5e9a4c03692c79e4538995ebb191 Mon Sep 17 00:00:00 2001 From: Xiyou Zhou Date: Mon, 15 May 2023 14:52:25 -0700 Subject: [PATCH] Update README.md Fix typo. --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 1f84946..4380ca5 100644 --- a/README.md +++ b/README.md @@ -126,7 +126,7 @@ Below is a short example demonstrating how to use the low-level API to tokenize >>> ctx = llama_cpp.llama_init_from_file(b"./models/7b/ggml-model.bin", params) >>> max_tokens = params.n_ctx # use ctypes arrays for array params ->>> tokens = (llama_cppp.llama_token * int(max_tokens))() +>>> tokens = (llama_cpp.llama_token * int(max_tokens))() >>> n_tokens = llama_cpp.llama_tokenize(ctx, b"Q: Name the planets in the solar system? A: ", tokens, max_tokens, add_bos=llama_cpp.c_bool(True)) >>> llama_cpp.llama_free(ctx) ```