From 6c31730e04a9aff407c52e5c3c55f7f413128153 Mon Sep 17 00:00:00 2001 From: Nate Sesti Date: Tue, 26 Sep 2023 10:19:31 -0700 Subject: docs: :memo: update references to Ollama to mention Linux support --- docs/docs/customization/models.md | 2 +- docs/docs/reference/Models/ollama.md | 3 +-- 2 files changed, 2 insertions(+), 3 deletions(-) (limited to 'docs') diff --git a/docs/docs/customization/models.md b/docs/docs/customization/models.md index cebb0667..ba3f6424 100644 --- a/docs/docs/customization/models.md +++ b/docs/docs/customization/models.md @@ -10,7 +10,7 @@ Commercial Models Local Models -- [Ollama](../reference/Models/ollama.md) - If you have a Mac, Ollama is the simplest way to run open-source models like Code Llama. +- [Ollama](../reference/Models/ollama.md) - If you are on Mac or Linux, Ollama is the simplest way to run open-source models like Code Llama. - [OpenAI](../reference/Models/openai.md) - If you have access to an OpenAI-compatible server (e.g. llama-cpp-python, LocalAI, FastChat, TextGenWebUI, etc.), you can use the `OpenAI` class and just change the base URL. - [GGML](../reference/Models/ggml.md) - An alternative way to connect to OpenAI-compatible servers. Will use `aiohttp` directly instead of the `openai` Python package. - [LlamaCpp](../reference/Models/llamacpp.md) - Build llama.cpp from source and use its built-in API server. diff --git a/docs/docs/reference/Models/ollama.md b/docs/docs/reference/Models/ollama.md index 39257395..6388e8cc 100644 --- a/docs/docs/reference/Models/ollama.md +++ b/docs/docs/reference/Models/ollama.md @@ -2,7 +2,7 @@ import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; # Ollama -[Ollama](https://ollama.ai/) is a Mac application that makes it easy to locally run open-source models, including Llama-2. Download the app from the website, and it will walk you through setup in a couple of minutes. You can also read more in their [README](https://github.com/jmorganca/ollama). Continue can then be configured to use the `Ollama` LLM class: +[Ollama](https://ollama.ai/) is an application for Mac and Linux that makes it easy to locally run open-source models, including Llama-2. Download the app from the website, and it will walk you through setup in a couple of minutes. You can also read more in their [README](https://github.com/jmorganca/ollama). Continue can then be configured to use the `Ollama` LLM class: ```python from continuedev.src.continuedev.libs.llm.ollama import Ollama @@ -21,7 +21,6 @@ config = ContinueConfig( - ### Inherited Properties -- cgit v1.2.3-70-g09d2