From b17c131928eb862cf32c8159fb7d1e003b69e9cd Mon Sep 17 00:00:00 2001 From: Nate Sesti Date: Sun, 27 Aug 2023 23:46:47 -0700 Subject: docs: :memo: update replicate documentation to reflect codellama --- docs/docs/customization.md | 2 +- docs/docs/walkthroughs/codellama.md | 3 ++- 2 files changed, 3 insertions(+), 2 deletions(-) diff --git a/docs/docs/customization.md b/docs/docs/customization.md index 3240c185..0a91b03b 100644 --- a/docs/docs/customization.md +++ b/docs/docs/customization.md @@ -129,7 +129,7 @@ config = ContinueConfig( ... models=Models( default=ReplicateLLM( - model="replicate/llama-2-70b-chat:58d078176e02c219e11eb4da5a02a7830a283b14cf8f94537af893ccff5ee781", + model="replicate/codellama-13b-instruct:da5676342de1a5a335b848383af297f592b816b950a43d251a0a9edd0113604b", api_key="my-replicate-api-key") ) ) diff --git a/docs/docs/walkthroughs/codellama.md b/docs/docs/walkthroughs/codellama.md index fdede918..68e99948 100644 --- a/docs/docs/walkthroughs/codellama.md +++ b/docs/docs/walkthroughs/codellama.md @@ -31,6 +31,7 @@ config = ContinueConfig( 1. Download Ollama [here](https://ollama.ai/) (it should walk you through the rest of these steps) 2. Open a terminal and run `ollama pull codellama`\* 3. Change your Continue config file to look like this: + ```python from continuedev.src.continuedev.libs.llm.ollama import Ollama @@ -59,7 +60,7 @@ config = ContinueConfig( ... models=Models( default=ReplicateLLM( - model="", + model="replicate/codellama-13b-instruct:da5676342de1a5a335b848383af297f592b816b950a43d251a0a9edd0113604b", api_key="") ) ) -- cgit v1.2.3-70-g09d2