diff options
author | Nate Sesti <sestinj@gmail.com> | 2023-09-26 10:19:31 -0700 |
---|---|---|
committer | Nate Sesti <sestinj@gmail.com> | 2023-09-26 10:19:31 -0700 |
commit | 6c31730e04a9aff407c52e5c3c55f7f413128153 (patch) | |
tree | a32ea58b4b359bed19c38b715f6f9ab3cdf0ea73 | |
parent | a94b675aa1634b356648a731283f74d761508b54 (diff) | |
download | sncontinue-6c31730e04a9aff407c52e5c3c55f7f413128153.tar.gz sncontinue-6c31730e04a9aff407c52e5c3c55f7f413128153.tar.bz2 sncontinue-6c31730e04a9aff407c52e5c3c55f7f413128153.zip |
docs: :memo: update references to Ollama to mention Linux support
-rw-r--r-- | continuedev/src/continuedev/libs/llm/ollama.py | 2 | ||||
-rw-r--r-- | continuedev/src/continuedev/plugins/steps/setup_model.py | 2 | ||||
-rw-r--r-- | docs/docs/customization/models.md | 2 | ||||
-rw-r--r-- | docs/docs/reference/Models/ollama.md | 3 | ||||
-rw-r--r-- | extension/react-app/src/pages/models.tsx | 2 |
5 files changed, 5 insertions, 6 deletions
diff --git a/continuedev/src/continuedev/libs/llm/ollama.py b/continuedev/src/continuedev/libs/llm/ollama.py index 19d48a2f..ee6ab540 100644 --- a/continuedev/src/continuedev/libs/llm/ollama.py +++ b/continuedev/src/continuedev/libs/llm/ollama.py @@ -12,7 +12,7 @@ from .prompts.edit import simplified_edit_prompt class Ollama(LLM): """ - [Ollama](https://ollama.ai/) is a Mac application that makes it easy to locally run open-source models, including Llama-2. Download the app from the website, and it will walk you through setup in a couple of minutes. You can also read more in their [README](https://github.com/jmorganca/ollama). Continue can then be configured to use the `Ollama` LLM class: + [Ollama](https://ollama.ai/) is an application for Mac and Linux that makes it easy to locally run open-source models, including Llama-2. Download the app from the website, and it will walk you through setup in a couple of minutes. You can also read more in their [README](https://github.com/jmorganca/ollama). Continue can then be configured to use the `Ollama` LLM class: ```python from continuedev.src.continuedev.libs.llm.ollama import Ollama diff --git a/continuedev/src/continuedev/plugins/steps/setup_model.py b/continuedev/src/continuedev/plugins/steps/setup_model.py index 83e616b0..f29bd51e 100644 --- a/continuedev/src/continuedev/plugins/steps/setup_model.py +++ b/continuedev/src/continuedev/plugins/steps/setup_model.py @@ -9,7 +9,7 @@ MODEL_CLASS_TO_MESSAGE = { "OpenAIFreeTrial": "To get started with OpenAI models, obtain your OpenAI API key from [here](https://platform.openai.com/account/api-keys) and paste it into the `api_key` field at config.models.default.api_key in `config.py`. Then reload the VS Code window for changes to take effect.", "AnthropicLLM": "To get started with Anthropic, you first need to sign up for the beta [here](https://claude.ai/login) to obtain an API key. Once you have the key, paste it into the `api_key` field at config.models.default.api_key in `config.py`. Then reload the VS Code window for changes to take effect.", "ReplicateLLM": "To get started with Replicate, sign up to obtain an API key [here](https://replicate.ai/), then paste it into the `api_key` field at config.models.default.api_key in `config.py`.", - "Ollama": "To get started with Ollama, download the Mac app from [ollama.ai](https://ollama.ai/). Once it is downloaded, be sure to pull at least one model and use its name in the model field in config.py (e.g. `model='codellama'`).", + "Ollama": "To get started with Ollama, download the app from [ollama.ai](https://ollama.ai/). Once it is downloaded, be sure to pull at least one model and use its name in the model field in config.py (e.g. `model='codellama'`).", "GGML": "GGML models can be run locally using the `llama-cpp-python` library. To learn how to set up a local llama-cpp-python server, read [here](https://github.com/continuedev/ggml-server-example). Once it is started on port 8000, you're all set!", "TogetherLLM": "To get started using models from Together, first obtain your Together API key from [here](https://together.ai). Paste it into the `api_key` field at config.models.default.api_key in `config.py`. Then, on their models page, press 'start' on the model of your choice and make sure the `model=` parameter in the config file for the `TogetherLLM` class reflects the name of this model. Finally, reload the VS Code window for changes to take effect.", "LlamaCpp": "To get started with this model, clone the [`llama.cpp` repo](https://github.com/ggerganov/llama.cpp) and follow the instructions to set up the server [here](https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md#build). Any of the parameters described in the README can be passed to the `llama_cpp_args` field in the `LlamaCpp` class in `config.py`.", diff --git a/docs/docs/customization/models.md b/docs/docs/customization/models.md index cebb0667..ba3f6424 100644 --- a/docs/docs/customization/models.md +++ b/docs/docs/customization/models.md @@ -10,7 +10,7 @@ Commercial Models Local Models -- [Ollama](../reference/Models/ollama.md) - If you have a Mac, Ollama is the simplest way to run open-source models like Code Llama. +- [Ollama](../reference/Models/ollama.md) - If you are on Mac or Linux, Ollama is the simplest way to run open-source models like Code Llama. - [OpenAI](../reference/Models/openai.md) - If you have access to an OpenAI-compatible server (e.g. llama-cpp-python, LocalAI, FastChat, TextGenWebUI, etc.), you can use the `OpenAI` class and just change the base URL. - [GGML](../reference/Models/ggml.md) - An alternative way to connect to OpenAI-compatible servers. Will use `aiohttp` directly instead of the `openai` Python package. - [LlamaCpp](../reference/Models/llamacpp.md) - Build llama.cpp from source and use its built-in API server. diff --git a/docs/docs/reference/Models/ollama.md b/docs/docs/reference/Models/ollama.md index 39257395..6388e8cc 100644 --- a/docs/docs/reference/Models/ollama.md +++ b/docs/docs/reference/Models/ollama.md @@ -2,7 +2,7 @@ import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; # Ollama -[Ollama](https://ollama.ai/) is a Mac application that makes it easy to locally run open-source models, including Llama-2. Download the app from the website, and it will walk you through setup in a couple of minutes. You can also read more in their [README](https://github.com/jmorganca/ollama). Continue can then be configured to use the `Ollama` LLM class: +[Ollama](https://ollama.ai/) is an application for Mac and Linux that makes it easy to locally run open-source models, including Llama-2. Download the app from the website, and it will walk you through setup in a couple of minutes. You can also read more in their [README](https://github.com/jmorganca/ollama). Continue can then be configured to use the `Ollama` LLM class: ```python from continuedev.src.continuedev.libs.llm.ollama import Ollama @@ -21,7 +21,6 @@ config = ContinueConfig( <ClassPropertyRef name='server_url' details='{"title": "Server Url", "description": "URL of the Ollama server", "default": "http://localhost:11434", "type": "string"}' required={false} default="http://localhost:11434"/> - ### Inherited Properties <ClassPropertyRef name='title' details='{"title": "Title", "description": "A title that will identify this model in the model selection dropdown", "type": "string"}' required={false} default=""/> diff --git a/extension/react-app/src/pages/models.tsx b/extension/react-app/src/pages/models.tsx index ea8e28d6..c20d820c 100644 --- a/extension/react-app/src/pages/models.tsx +++ b/extension/react-app/src/pages/models.tsx @@ -35,7 +35,7 @@ const MODEL_INFO: ModelInfo[] = [ title: "Ollama", class: "Ollama", description: - "One of the fastest ways to get started with local models on Mac", + "One of the fastest ways to get started with local models on Mac or Linux", args: { model: "codellama", title: "Ollama", |