diff options
| author | Nate Sesti <sestinj@gmail.com> | 2023-09-07 00:57:28 -0700 | 
|---|---|---|
| committer | Nate Sesti <sestinj@gmail.com> | 2023-09-07 00:57:28 -0700 | 
| commit | 5c6609ab5fa3a69cd0e3e8e61df643fcce1ecb47 (patch) | |
| tree | a5c6a89db6b8cd41835280500aa6c7a7736968f9 /continuedev/src | |
| parent | 44070ac1851ad19ebdf2c943fe06083da23c9141 (diff) | |
| download | sncontinue-5c6609ab5fa3a69cd0e3e8e61df643fcce1ecb47.tar.gz sncontinue-5c6609ab5fa3a69cd0e3e8e61df643fcce1ecb47.tar.bz2 sncontinue-5c6609ab5fa3a69cd0e3e8e61df643fcce1ecb47.zip  | |
fix: :bug: templating fix for queued LLM
Diffstat (limited to 'continuedev/src')
| -rw-r--r-- | continuedev/src/continuedev/libs/llm/ggml.py | 6 | ||||
| -rw-r--r-- | continuedev/src/continuedev/libs/llm/queued.py | 3 | 
2 files changed, 7 insertions, 2 deletions
diff --git a/continuedev/src/continuedev/libs/llm/ggml.py b/continuedev/src/continuedev/libs/llm/ggml.py index 3fbfdeed..dd1bdec2 100644 --- a/continuedev/src/continuedev/libs/llm/ggml.py +++ b/continuedev/src/continuedev/libs/llm/ggml.py @@ -1,6 +1,6 @@  import json  import ssl -from typing import Any, Coroutine, List, Optional +from typing import Any, Callable, Coroutine, Dict, List, Optional  import aiohttp @@ -19,7 +19,9 @@ class GGML(LLM):      ca_bundle_path: str = None      model: str = "ggml" -    template_messages = llama2_template_messages +    template_messages: Optional[ +        Callable[[List[Dict[str, str]]], str] +    ] = llama2_template_messages      prompt_templates = {          "edit": simplified_edit_prompt, diff --git a/continuedev/src/continuedev/libs/llm/queued.py b/continuedev/src/continuedev/libs/llm/queued.py index 9e6e0180..6dbaaa64 100644 --- a/continuedev/src/continuedev/libs/llm/queued.py +++ b/continuedev/src/continuedev/libs/llm/queued.py @@ -19,6 +19,9 @@ class QueuedLLM(LLM):          await self.llm.start(*args, **kwargs)          self._lock = asyncio.Lock()          self.model = self.llm.model +        self.template_messages = self.llm.template_messages +        self.prompt_templates = self.llm.prompt_templates +        self.context_length = self.llm.context_length      async def stop(self):          await self.llm.stop()  | 
