From 1e3c8adabba561eeef124144f3a2ef36d26334b4 Mon Sep 17 00:00:00 2001 From: Nate Sesti Date: Sun, 24 Sep 2023 17:46:33 -0700 Subject: feat: :fire: fix duplication in reference --- docs/docs/reference/Models/anthropic.md | 39 ------------------- docs/docs/reference/Models/hf_inference_api.md | 42 --------------------- docs/docs/reference/Models/hf_tgi.md | 27 ------------- docs/docs/reference/Models/maybe_proxy_openai.md | 47 ----------------------- docs/docs/reference/Models/openai_free_trial.md | 48 ------------------------ docs/docs/reference/Models/openaifreetrial.md | 1 + docs/docs/reference/Models/queued.md | 40 -------------------- docs/docs/reference/Models/replicate.md | 42 --------------------- docs/docs/reference/Models/text_gen_interface.md | 41 -------------------- docs/docs/reference/Models/together.md | 42 --------------------- docs/docs/reference/config.md | 2 + 11 files changed, 3 insertions(+), 368 deletions(-) delete mode 100644 docs/docs/reference/Models/anthropic.md delete mode 100644 docs/docs/reference/Models/hf_inference_api.md delete mode 100644 docs/docs/reference/Models/hf_tgi.md delete mode 100644 docs/docs/reference/Models/maybe_proxy_openai.md delete mode 100644 docs/docs/reference/Models/openai_free_trial.md delete mode 100644 docs/docs/reference/Models/queued.md delete mode 100644 docs/docs/reference/Models/replicate.md delete mode 100644 docs/docs/reference/Models/text_gen_interface.md delete mode 100644 docs/docs/reference/Models/together.md diff --git a/docs/docs/reference/Models/anthropic.md b/docs/docs/reference/Models/anthropic.md deleted file mode 100644 index 128b706d..00000000 --- a/docs/docs/reference/Models/anthropic.md +++ /dev/null @@ -1,39 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# AnthropicLLM - -Import the `AnthropicLLM` class and set it as the default model: - -```python -from continuedev.src.continuedev.libs.llm.anthropic import AnthropicLLM - -config = ContinueConfig( - ... - models=Models( - default=AnthropicLLM(api_key="", model="claude-2") - ) -) -``` - -Claude 2 is not yet publicly released. You can request early access [here](https://www.anthropic.com/earlyaccess). - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/anthropic.py) - -## Properties - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/hf_inference_api.md b/docs/docs/reference/Models/hf_inference_api.md deleted file mode 100644 index 560309f2..00000000 --- a/docs/docs/reference/Models/hf_inference_api.md +++ /dev/null @@ -1,42 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# HuggingFaceInferenceAPI - -Hugging Face Inference API is a great option for newly released language models. Sign up for an account and add billing [here](https://huggingface.co/settings/billing), access the Inference Endpoints [here](https://ui.endpoints.huggingface.co), click on “New endpoint”, and fill out the form (e.g. select a model like [WizardCoder-Python-34B-V1.0](https://huggingface.co/WizardLM/WizardCoder-Python-34B-V1.0)), and then deploy your model by clicking “Create Endpoint”. Change `~/.continue/config.py` to look like this: - -```python -from continuedev.src.continuedev.core.models import Models -from continuedev.src.continuedev.libs.llm.hf_inference_api import HuggingFaceInferenceAPI - -config = ContinueConfig( - ... - models=Models( - default=HuggingFaceInferenceAPI( - endpoint_url: "", - hf_token: "", - ) -) -``` - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/hf_inference_api.py) - -## Properties - - - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/hf_tgi.md b/docs/docs/reference/Models/hf_tgi.md deleted file mode 100644 index 2cee9fe1..00000000 --- a/docs/docs/reference/Models/hf_tgi.md +++ /dev/null @@ -1,27 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# HuggingFaceTGI - - - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/hf_tgi.py) - -## Properties - - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/maybe_proxy_openai.md b/docs/docs/reference/Models/maybe_proxy_openai.md deleted file mode 100644 index 055054fd..00000000 --- a/docs/docs/reference/Models/maybe_proxy_openai.md +++ /dev/null @@ -1,47 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# OpenAIFreeTrial - -With the `OpenAIFreeTrial` `LLM`, new users can try out Continue with GPT-4 using a proxy server that securely makes calls to OpenAI using our API key. Continue should just work the first time you install the extension in VS Code. - -Once you are using Continue regularly though, you will need to add an OpenAI API key that has access to GPT-4 by following these steps: - -1. Copy your API key from https://platform.openai.com/account/api-keys -2. Open `~/.continue/config.py`. You can do this by using the '/config' command in Continue -3. Change the default LLMs to look like this: - -```python -API_KEY = "" -config = ContinueConfig( - ... - models=Models( - default=OpenAIFreeTrial(model="gpt-4", api_key=API_KEY), - medium=OpenAIFreeTrial(model="gpt-3.5-turbo", api_key=API_KEY) - ) -) -``` - -The `OpenAIFreeTrial` class will automatically switch to using your API key instead of ours. If you'd like to explicitly use one or the other, you can use the `ProxyServer` or `OpenAI` classes instead. - -These classes support any models available through the OpenAI API, assuming your API key has access, including "gpt-4", "gpt-3.5-turbo", "gpt-3.5-turbo-16k", and "gpt-4-32k". - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/openai_free_trial.py) - -## Properties - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/openai_free_trial.md b/docs/docs/reference/Models/openai_free_trial.md deleted file mode 100644 index cd510aa8..00000000 --- a/docs/docs/reference/Models/openai_free_trial.md +++ /dev/null @@ -1,48 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# OpenAIFreeTrial - -With the `OpenAIFreeTrial` `LLM`, new users can try out Continue with GPT-4 using a proxy server that securely makes calls to OpenAI using our API key. Continue should just work the first time you install the extension in VS Code. - -Once you are using Continue regularly though, you will need to add an OpenAI API key that has access to GPT-4 by following these steps: - -1. Copy your API key from https://platform.openai.com/account/api-keys -2. Open `~/.continue/config.py`. You can do this by using the '/config' command in Continue -3. Change the default LLMs to look like this: - -```python -API_KEY = "" -config = ContinueConfig( - ... - models=Models( - default=OpenAIFreeTrial(model="gpt-4", api_key=API_KEY), - medium=OpenAIFreeTrial(model="gpt-3.5-turbo", api_key=API_KEY) - ) -) -``` - -The `OpenAIFreeTrial` class will automatically switch to using your API key instead of ours. If you'd like to explicitly use one or the other, you can use the `ProxyServer` or `OpenAI` classes instead. - -These classes support any models available through the OpenAI API, assuming your API key has access, including "gpt-4", "gpt-3.5-turbo", "gpt-3.5-turbo-16k", and "gpt-4-32k". - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/openai_free_trial.py) - -## Properties - - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/openaifreetrial.md b/docs/docs/reference/Models/openaifreetrial.md index a9efa6cc..99c21689 100644 --- a/docs/docs/reference/Models/openaifreetrial.md +++ b/docs/docs/reference/Models/openaifreetrial.md @@ -31,6 +31,7 @@ These classes support any models available through the OpenAI API, assuming your + ### Inherited Properties diff --git a/docs/docs/reference/Models/queued.md b/docs/docs/reference/Models/queued.md deleted file mode 100644 index 06942e3e..00000000 --- a/docs/docs/reference/Models/queued.md +++ /dev/null @@ -1,40 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# QueuedLLM - -QueuedLLM exists to make up for LLM servers that cannot handle multiple requests at once. It uses a lock to ensure that only one request is being processed at a time. - -If you are already using another LLM class and are experiencing this problem, you can just wrap it with the QueuedLLM class like this: - -```python -from continuedev.src.continuedev.libs.llm.queued import QueuedLLM - -config = ContinueConfig( - ... - models=Models( - default=QueuedLLM(llm=) - ) -) -``` - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/queued.py) - -## Properties - - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/replicate.md b/docs/docs/reference/Models/replicate.md deleted file mode 100644 index 879459e0..00000000 --- a/docs/docs/reference/Models/replicate.md +++ /dev/null @@ -1,42 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# ReplicateLLM - -Replicate is a great option for newly released language models or models that you've deployed through their platform. Sign up for an account [here](https://replicate.ai/), copy your API key, and then select any model from the [Replicate Streaming List](https://replicate.com/collections/streaming-language-models). Change `~/.continue/config.py` to look like this: - -```python -from continuedev.src.continuedev.core.models import Models -from continuedev.src.continuedev.libs.llm.replicate import ReplicateLLM - -config = ContinueConfig( - ... - models=Models( - default=ReplicateLLM( - model="replicate/codellama-13b-instruct:da5676342de1a5a335b848383af297f592b816b950a43d251a0a9edd0113604b", - api_key="my-replicate-api-key") - ) -) -``` - -If you don't specify the `model` parameter, it will default to `replicate/llama-2-70b-chat:58d078176e02c219e11eb4da5a02a7830a283b14cf8f94537af893ccff5ee781`. - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/replicate.py) - -## Properties - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/text_gen_interface.md b/docs/docs/reference/Models/text_gen_interface.md deleted file mode 100644 index bb8dce1d..00000000 --- a/docs/docs/reference/Models/text_gen_interface.md +++ /dev/null @@ -1,41 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# TextGenUI - -TextGenUI is a comprehensive, open-source language model UI and local server. You can set it up with an OpenAI-compatible server plugin, but if for some reason that doesn't work, you can use this class like so: - -```python -from continuedev.src.continuedev.libs.llm.text_gen_interface import TextGenUI - -config = ContinueConfig( - ... - models=Models( - default=TextGenUI( - model="", - ) - ) -) -``` - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/text_gen_interface.py) - -## Properties - - - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/Models/together.md b/docs/docs/reference/Models/together.md deleted file mode 100644 index 3718f046..00000000 --- a/docs/docs/reference/Models/together.md +++ /dev/null @@ -1,42 +0,0 @@ -import ClassPropertyRef from '@site/src/components/ClassPropertyRef.tsx'; - -# TogetherLLM - -The Together API is a cloud platform for running large AI models. You can sign up [here](https://api.together.xyz/signup), copy your API key on the initial welcome screen, and then hit the play button on any model from the [Together Models list](https://docs.together.ai/docs/models-inference). Change `~/.continue/config.py` to look like this: - -```python -from continuedev.src.continuedev.core.models import Models -from continuedev.src.continuedev.libs.llm.together import TogetherLLM - -config = ContinueConfig( - ... - models=Models( - default=TogetherLLM( - api_key="", - model="togethercomputer/llama-2-13b-chat" - ) - ) -) -``` - -[View the source](https://github.com/continuedev/continue/tree/main/continuedev/src/continuedev/libs/llm/together.py) - -## Properties - - - - -### Inherited Properties - - - - - - - - - - - - - diff --git a/docs/docs/reference/config.md b/docs/docs/reference/config.md index 1f683ed2..60d5b73e 100644 --- a/docs/docs/reference/config.md +++ b/docs/docs/reference/config.md @@ -23,4 +23,6 @@ Continue can be deeply customized by editing the `ContinueConfig` object in `~/. + ### Inherited Properties + -- cgit v1.2.3-70-g09d2