diff options
author | Nate Sesti <sestinj@gmail.com> | 2023-07-30 16:26:31 -0700 |
---|---|---|
committer | Nate Sesti <sestinj@gmail.com> | 2023-07-30 16:26:31 -0700 |
commit | 6ead69fb71ea01e8b0ab6964d17c5dd058244883 (patch) | |
tree | 8e15b75bb09ab28560f2fba2387678f51a1ff743 /continuedev | |
parent | e37996002d848fc71c82375199dc9a704f2c9b05 (diff) | |
download | sncontinue-6ead69fb71ea01e8b0ab6964d17c5dd058244883.tar.gz sncontinue-6ead69fb71ea01e8b0ab6964d17c5dd058244883.tar.bz2 sncontinue-6ead69fb71ea01e8b0ab6964d17c5dd058244883.zip |
refactor: :zap: turn GGML's aiohttp.AsyncSession into an instance attribute
Diffstat (limited to 'continuedev')
-rw-r--r-- | continuedev/src/continuedev/libs/constants/default_config.py.txt | 3 | ||||
-rw-r--r-- | continuedev/src/continuedev/libs/llm/ggml.py | 10 |
2 files changed, 8 insertions, 5 deletions
diff --git a/continuedev/src/continuedev/libs/constants/default_config.py.txt b/continuedev/src/continuedev/libs/constants/default_config.py.txt index 7cd2226a..7c7f495e 100644 --- a/continuedev/src/continuedev/libs/constants/default_config.py.txt +++ b/continuedev/src/continuedev/libs/constants/default_config.py.txt @@ -45,7 +45,8 @@ config = ContinueConfig( allow_anonymous_telemetry=True, models=Models( - default=MaybeProxyOpenAI("gpt4") + default=MaybeProxyOpenAI("gpt4"), + medium=MaybeProxyOpenAI("gpt-3.5-turbo") ) # Set a system message with information that the LLM should always keep in mind diff --git a/continuedev/src/continuedev/libs/llm/ggml.py b/continuedev/src/continuedev/libs/llm/ggml.py index 4bcf7e54..990f35bc 100644 --- a/continuedev/src/continuedev/libs/llm/ggml.py +++ b/continuedev/src/continuedev/libs/llm/ggml.py @@ -13,11 +13,13 @@ SERVER_URL = "http://localhost:8000" class GGML(LLM): + _client_session: aiohttp.ClientSession + def __init__(self, system_message: str = None): self.system_message = system_message async def start(self, **kwargs): - pass + self._client_session = aiohttp.ClientSession() async def stop(self): pass @@ -43,7 +45,7 @@ class GGML(LLM): self.name, with_history, args["max_tokens"], prompt, functions=args.get("functions", None), system_message=self.system_message) # TODO move to single self.session variable (proxy setting etc) - async with aiohttp.ClientSession() as session: + async with self._client_session as session: async with session.post(f"{SERVER_URL}/v1/completions", json={ "messages": messages, **args @@ -61,7 +63,7 @@ class GGML(LLM): self.name, messages, args["max_tokens"], None, functions=args.get("functions", None), system_message=self.system_message) args["stream"] = True - async with aiohttp.ClientSession() as session: + async with self._client_session as session: async with session.post(f"{SERVER_URL}/v1/chat/completions", json={ "messages": messages, **args @@ -83,7 +85,7 @@ class GGML(LLM): async def complete(self, prompt: str, with_history: List[ChatMessage] = None, **kwargs) -> Coroutine[Any, Any, str]: args = {**self.default_args, **kwargs} - async with aiohttp.ClientSession() as session: + async with self._client_session as session: async with session.post(f"{SERVER_URL}/v1/completions", json={ "messages": compile_chat_messages(args["model"], with_history, args["max_tokens"], prompt, functions=None, system_message=self.system_message), **args |