From 12199bcfa85cac6854953381976dc9956b4c0cf7 Mon Sep 17 00:00:00 2001 From: Jeffrey Morgan Date: Fri, 7 Jul 2023 15:21:40 -0400 Subject: [PATCH] update download links to the releases page until we have a better download url --- README.md | 2 +- web/app/page.tsx | 13 +++++-------- web/package-lock.json | 4 ++-- 3 files changed, 8 insertions(+), 11 deletions(-) diff --git a/README.md b/README.md index 99834a42..0c57406b 100644 --- a/README.md +++ b/README.md @@ -16,7 +16,7 @@ Run large language models with `llama.cpp`. ## Install -- Download for macOS +- [Download](https://github.com/jmorganca/ollama/releases/latest) for macOS - Download for Windows (coming soon) - Docker: `docker run -p 11434:11434 ollama/ollama` diff --git a/web/app/page.tsx b/web/app/page.tsx index c28f8871..2a3476f9 100644 --- a/web/app/page.tsx +++ b/web/app/page.tsx @@ -5,18 +5,15 @@ export default async function Home() {

ollama

-

+

Ollama {' '} - is a tool for running large language models. -
-
- Get started with Ollama using pip: + is a tool for running large language models. The latest version is available for download{' '} + + here. +

-
-          pip install ollama
-        

Example models you can try running:

diff --git a/web/package-lock.json b/web/package-lock.json index ca8bdeb8..02d8b622 100644 --- a/web/package-lock.json +++ b/web/package-lock.json @@ -1,12 +1,12 @@ { "name": "web", - "version": "0.1.0", + "version": "0.0.0", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "web", - "version": "0.1.0", + "version": "0.0.0", "dependencies": { "@octokit/rest": "^19.0.13", "@types/node": "20.4.0",