summaryrefslogtreecommitdiff
path: root/continuedev
diff options
context:
space:
mode:
Diffstat (limited to 'continuedev')
-rw-r--r--continuedev/src/continuedev/libs/llm/__init__.py4
-rw-r--r--continuedev/src/continuedev/libs/llm/ggml.py4
-rw-r--r--continuedev/src/continuedev/libs/llm/hf_tgi.py4
-rw-r--r--continuedev/src/continuedev/libs/llm/llamacpp.py4
-rw-r--r--continuedev/src/continuedev/libs/llm/maybe_proxy_openai.py4
-rw-r--r--continuedev/src/continuedev/libs/llm/openai.py4
-rw-r--r--continuedev/src/continuedev/libs/llm/proxy_server.py2
-rw-r--r--continuedev/src/continuedev/libs/llm/text_gen_interface.py3
-rw-r--r--continuedev/src/continuedev/libs/llm/together.py4
9 files changed, 7 insertions, 26 deletions
diff --git a/continuedev/src/continuedev/libs/llm/__init__.py b/continuedev/src/continuedev/libs/llm/__init__.py
index 11d81b3f..8c61ba43 100644
--- a/continuedev/src/continuedev/libs/llm/__init__.py
+++ b/continuedev/src/continuedev/libs/llm/__init__.py
@@ -72,6 +72,9 @@ class LLM(ContinueBaseModel):
300,
description="Set the timeout for each request to the LLM. If you are running a local LLM that takes a while to respond, you might want to set this to avoid timeouts.",
)
+ verify_ssl: Optional[bool] = Field(
+ None, description="Whether to verify SSL certificates for requests."
+ )
prompt_templates: dict = Field(
{},
description='A dictionary of prompt templates that can be used to customize the behavior of the LLM in certain situations. For example, set the "edit" key in order to change the prompt that is used for the /edit slash command. Each value in the dictionary is a string templated in mustache syntax, and filled in at runtime with the variables specific to the situation. See the documentation for more information.',
@@ -120,6 +123,7 @@ class LLM(ContinueBaseModel):
"description": "A function that is called upon every prompt and completion, by default to log to the file which can be viewed by clicking on the magnifying glass."
},
"api_key": {"description": "The API key for the LLM provider."},
+ "verify_ssl": {"description": "Whether to verify SSL certificates for requests."}
}
def dict(self, **kwargs):
diff --git a/continuedev/src/continuedev/libs/llm/ggml.py b/continuedev/src/continuedev/libs/llm/ggml.py
index 0ab43703..f20d8b45 100644
--- a/continuedev/src/continuedev/libs/llm/ggml.py
+++ b/continuedev/src/continuedev/libs/llm/ggml.py
@@ -38,10 +38,6 @@ class GGML(LLM):
"http://localhost:8000",
description="URL of the OpenAI-compatible server where the model is being served",
)
- verify_ssl: Optional[bool] = Field(
- None,
- description="Whether SSL certificates should be verified when making the HTTP request",
- )
proxy: Optional[str] = Field(
None,
description="Proxy URL to use when making the HTTP request",
diff --git a/continuedev/src/continuedev/libs/llm/hf_tgi.py b/continuedev/src/continuedev/libs/llm/hf_tgi.py
index 6b7f21e7..168ef025 100644
--- a/continuedev/src/continuedev/libs/llm/hf_tgi.py
+++ b/continuedev/src/continuedev/libs/llm/hf_tgi.py
@@ -15,10 +15,6 @@ class HuggingFaceTGI(LLM):
server_url: str = Field(
"http://localhost:8080", description="URL of your TGI server"
)
- verify_ssl: Optional[bool] = Field(
- None,
- description="Whether SSL certificates should be verified when making the HTTP request",
- )
template_messages: Callable[[List[ChatMessage]], str] = code_llama_template_messages
diff --git a/continuedev/src/continuedev/libs/llm/llamacpp.py b/continuedev/src/continuedev/libs/llm/llamacpp.py
index 10dfcad8..60d9961f 100644
--- a/continuedev/src/continuedev/libs/llm/llamacpp.py
+++ b/continuedev/src/continuedev/libs/llm/llamacpp.py
@@ -35,10 +35,6 @@ class LlamaCpp(LLM):
model: str = "llamacpp"
server_url: str = Field("http://localhost:8080", description="URL of the server")
- verify_ssl: Optional[bool] = Field(
- None,
- description="Whether SSL certificates should be verified when making the HTTP request",
- )
llama_cpp_args: Dict[str, Any] = Field(
{"stop": ["[INST]"]},
diff --git a/continuedev/src/continuedev/libs/llm/maybe_proxy_openai.py b/continuedev/src/continuedev/libs/llm/maybe_proxy_openai.py
index 69def48e..5814f2aa 100644
--- a/continuedev/src/continuedev/libs/llm/maybe_proxy_openai.py
+++ b/continuedev/src/continuedev/libs/llm/maybe_proxy_openai.py
@@ -45,9 +45,9 @@ class MaybeProxyOpenAI(LLM):
):
await super().start(write_log=write_log, unique_id=unique_id)
if self.api_key is None or self.api_key.strip() == "":
- self.llm = ProxyServer(model=self.model)
+ self.llm = ProxyServer(model=self.model, verify_ssl=self.verify_ssl)
else:
- self.llm = OpenAI(api_key=self.api_key, model=self.model)
+ self.llm = OpenAI(api_key=self.api_key, model=self.model, verify_ssl=self.verify_ssl)
await self.llm.start(write_log=write_log, unique_id=unique_id)
diff --git a/continuedev/src/continuedev/libs/llm/openai.py b/continuedev/src/continuedev/libs/llm/openai.py
index 744c07d2..c7ef9d95 100644
--- a/continuedev/src/continuedev/libs/llm/openai.py
+++ b/continuedev/src/continuedev/libs/llm/openai.py
@@ -60,10 +60,6 @@ class OpenAI(LLM):
description="OpenAI API key",
)
- verify_ssl: Optional[bool] = Field(
- None, description="Whether to verify SSL certificates for requests."
- )
-
ca_bundle_path: Optional[str] = Field(
None, description="Path to CA bundle to use for requests."
)
diff --git a/continuedev/src/continuedev/libs/llm/proxy_server.py b/continuedev/src/continuedev/libs/llm/proxy_server.py
index d62fafa7..032464be 100644
--- a/continuedev/src/continuedev/libs/llm/proxy_server.py
+++ b/continuedev/src/continuedev/libs/llm/proxy_server.py
@@ -36,7 +36,7 @@ class ProxyServer(LLM):
):
await super().start(**kwargs)
self._client_session = aiohttp.ClientSession(
- connector=aiohttp.TCPConnector(ssl_context=ssl_context),
+ connector=aiohttp.TCPConnector(ssl_context=ssl_context, verify_ssl=self.verify_ssl),
timeout=aiohttp.ClientTimeout(total=self.timeout),
)
self.context_length = MAX_TOKENS_FOR_MODEL[self.model]
diff --git a/continuedev/src/continuedev/libs/llm/text_gen_interface.py b/continuedev/src/continuedev/libs/llm/text_gen_interface.py
index 1090d7dd..28b2bfae 100644
--- a/continuedev/src/continuedev/libs/llm/text_gen_interface.py
+++ b/continuedev/src/continuedev/libs/llm/text_gen_interface.py
@@ -35,9 +35,6 @@ class TextGenUI(LLM):
"http://localhost:5005",
description="URL of your TextGenUI streaming server (separate from main server URL)",
)
- verify_ssl: Optional[bool] = Field(
- None, description="Whether to verify SSL certificates for requests."
- )
prompt_templates = {
"edit": simplest_edit_prompt,
diff --git a/continuedev/src/continuedev/libs/llm/together.py b/continuedev/src/continuedev/libs/llm/together.py
index a381abab..257f9a8f 100644
--- a/continuedev/src/continuedev/libs/llm/together.py
+++ b/continuedev/src/continuedev/libs/llm/together.py
@@ -38,10 +38,6 @@ class TogetherLLM(LLM):
"https://api.together.xyz",
description="The base URL for your Together API instance",
)
- verify_ssl: Optional[bool] = Field(
- None,
- description="Whether SSL certificates should be verified when making the HTTP request",
- )
_client_session: aiohttp.ClientSession = None