summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--continuedev/src/continuedev/libs/constants/default_config.py.txt3
-rw-r--r--continuedev/src/continuedev/libs/llm/ggml.py10
2 files changed, 8 insertions, 5 deletions
diff --git a/continuedev/src/continuedev/libs/constants/default_config.py.txt b/continuedev/src/continuedev/libs/constants/default_config.py.txt
index 7cd2226a..7c7f495e 100644
--- a/continuedev/src/continuedev/libs/constants/default_config.py.txt
+++ b/continuedev/src/continuedev/libs/constants/default_config.py.txt
@@ -45,7 +45,8 @@ config = ContinueConfig(
allow_anonymous_telemetry=True,
models=Models(
- default=MaybeProxyOpenAI("gpt4")
+ default=MaybeProxyOpenAI("gpt4"),
+ medium=MaybeProxyOpenAI("gpt-3.5-turbo")
)
# Set a system message with information that the LLM should always keep in mind
diff --git a/continuedev/src/continuedev/libs/llm/ggml.py b/continuedev/src/continuedev/libs/llm/ggml.py
index 4bcf7e54..990f35bc 100644
--- a/continuedev/src/continuedev/libs/llm/ggml.py
+++ b/continuedev/src/continuedev/libs/llm/ggml.py
@@ -13,11 +13,13 @@ SERVER_URL = "http://localhost:8000"
class GGML(LLM):
+ _client_session: aiohttp.ClientSession
+
def __init__(self, system_message: str = None):
self.system_message = system_message
async def start(self, **kwargs):
- pass
+ self._client_session = aiohttp.ClientSession()
async def stop(self):
pass
@@ -43,7 +45,7 @@ class GGML(LLM):
self.name, with_history, args["max_tokens"], prompt, functions=args.get("functions", None), system_message=self.system_message)
# TODO move to single self.session variable (proxy setting etc)
- async with aiohttp.ClientSession() as session:
+ async with self._client_session as session:
async with session.post(f"{SERVER_URL}/v1/completions", json={
"messages": messages,
**args
@@ -61,7 +63,7 @@ class GGML(LLM):
self.name, messages, args["max_tokens"], None, functions=args.get("functions", None), system_message=self.system_message)
args["stream"] = True
- async with aiohttp.ClientSession() as session:
+ async with self._client_session as session:
async with session.post(f"{SERVER_URL}/v1/chat/completions", json={
"messages": messages,
**args
@@ -83,7 +85,7 @@ class GGML(LLM):
async def complete(self, prompt: str, with_history: List[ChatMessage] = None, **kwargs) -> Coroutine[Any, Any, str]:
args = {**self.default_args, **kwargs}
- async with aiohttp.ClientSession() as session:
+ async with self._client_session as session:
async with session.post(f"{SERVER_URL}/v1/completions", json={
"messages": compile_chat_messages(args["model"], with_history, args["max_tokens"], prompt, functions=None, system_message=self.system_message),
**args