diff options
author | Nate Sesti <sestinj@gmail.com> | 2023-06-18 15:41:09 -0700 |
---|---|---|
committer | Nate Sesti <sestinj@gmail.com> | 2023-06-18 15:41:09 -0700 |
commit | 53b7fac40d572217c7167075d765ae7dfff56fbe (patch) | |
tree | a3885a6ffae121c5dd68747374be05c2abc1df03 /continuedev/src | |
parent | 9f4f8d48d51be2d140a722418d0a6c699c0f6cde (diff) | |
download | sncontinue-53b7fac40d572217c7167075d765ae7dfff56fbe.tar.gz sncontinue-53b7fac40d572217c7167075d765ae7dfff56fbe.tar.bz2 sncontinue-53b7fac40d572217c7167075d765ae7dfff56fbe.zip |
a number of things
Diffstat (limited to 'continuedev/src')
-rw-r--r-- | continuedev/src/continuedev/core/autopilot.py | 2 | ||||
-rw-r--r-- | continuedev/src/continuedev/libs/llm/openai.py | 4 | ||||
-rw-r--r-- | continuedev/src/continuedev/server/main.py | 5 | ||||
-rw-r--r-- | continuedev/src/continuedev/steps/chat.py | 2 | ||||
-rw-r--r-- | continuedev/src/continuedev/steps/react.py | 6 |
5 files changed, 11 insertions, 8 deletions
diff --git a/continuedev/src/continuedev/core/autopilot.py b/continuedev/src/continuedev/core/autopilot.py index 73f46a37..ee249c0b 100644 --- a/continuedev/src/continuedev/core/autopilot.py +++ b/continuedev/src/continuedev/core/autopilot.py @@ -120,7 +120,7 @@ class Autopilot(ContinueBaseModel): # If a parent step is deleted/cancelled, don't run this step last_depth = self._step_depth i = self.history.current_index - while i >= 0 and self.history.timeline[i].depth > last_depth: + while i >= 0 and self.history.timeline[i].depth == last_depth + 1: if self.history.timeline[i].deleted: return None last_depth = self.history.timeline[i].depth diff --git a/continuedev/src/continuedev/libs/llm/openai.py b/continuedev/src/continuedev/libs/llm/openai.py index 22c28b20..5d65eb22 100644 --- a/continuedev/src/continuedev/libs/llm/openai.py +++ b/continuedev/src/continuedev/libs/llm/openai.py @@ -83,14 +83,14 @@ class OpenAI(LLM): def with_system_message(self, system_message: Union[str, None]): return OpenAI(api_key=self.api_key, default_model=self.default_model, system_message=system_message) - def stream_chat(self, prompt, with_history: List[ChatMessage] = [], **kwargs) -> Generator[Union[Any, List, Dict], None, None]: + async def stream_chat(self, prompt, with_history: List[ChatMessage] = [], **kwargs) -> Generator[Union[Any, List, Dict], None, None]: self.completion_count += 1 args = {"max_tokens": DEFAULT_MAX_TOKENS, "temperature": 0.5, "top_p": 1, "frequency_penalty": 0, "presence_penalty": 0} | kwargs args["stream"] = True args["model"] = "gpt-3.5-turbo" - for chunk in openai.ChatCompletion.create( + async for chunk in await openai.ChatCompletion.acreate( messages=self.compile_chat_messages(with_history, prompt), **args, ): diff --git a/continuedev/src/continuedev/server/main.py b/continuedev/src/continuedev/server/main.py index b8bfe325..fb6ead02 100644 --- a/continuedev/src/continuedev/server/main.py +++ b/continuedev/src/continuedev/server/main.py @@ -40,10 +40,7 @@ args = parser.parse_args() def run_server(): - if os.path.exists("logging.yaml"): - uvicorn.run(app, host="0.0.0.0", port=args.port, log_level="info") - else: - uvicorn.run(app, host="0.0.0.0", port=args.port, log_level="info") + uvicorn.run(app, host="0.0.0.0", port=args.port) if __name__ == "__main__": diff --git a/continuedev/src/continuedev/steps/chat.py b/continuedev/src/continuedev/steps/chat.py index 90514ad6..fd7457d9 100644 --- a/continuedev/src/continuedev/steps/chat.py +++ b/continuedev/src/continuedev/steps/chat.py @@ -11,6 +11,8 @@ class SimpleChatStep(Step): async def run(self, sdk: ContinueSDK): self.description = f"```{self.user_input}```\n\n" + await sdk.update_ui() + async for chunk in sdk.models.default.stream_chat(self.user_input, with_history=await sdk.get_chat_context()): self.description += chunk await sdk.update_ui() diff --git a/continuedev/src/continuedev/steps/react.py b/continuedev/src/continuedev/steps/react.py index 4d310fc8..cddb8b42 100644 --- a/continuedev/src/continuedev/steps/react.py +++ b/continuedev/src/continuedev/steps/react.py @@ -10,7 +10,8 @@ class NLDecisionStep(Step): default_step: Union[Step, None] = None steps: List[Tuple[Step, str]] - hide: bool = True + hide: bool = False + name: str = "Deciding what to do next" async def run(self, sdk: ContinueSDK): step_descriptions = "\n".join([ @@ -36,4 +37,7 @@ class NLDecisionStep(Step): step_to_run = step_to_run or self.default_step or self.steps[0] + self.hide = True + await sdk.update_ui() + await sdk.run_step(step_to_run) |