diff options
Diffstat (limited to 'continuedev/src')
| -rw-r--r-- | continuedev/src/continuedev/core/autopilot.py | 2 | ||||
| -rw-r--r-- | continuedev/src/continuedev/libs/llm/openai.py | 4 | ||||
| -rw-r--r-- | continuedev/src/continuedev/server/main.py | 5 | ||||
| -rw-r--r-- | continuedev/src/continuedev/steps/chat.py | 2 | ||||
| -rw-r--r-- | continuedev/src/continuedev/steps/react.py | 6 | 
5 files changed, 11 insertions, 8 deletions
| diff --git a/continuedev/src/continuedev/core/autopilot.py b/continuedev/src/continuedev/core/autopilot.py index 73f46a37..ee249c0b 100644 --- a/continuedev/src/continuedev/core/autopilot.py +++ b/continuedev/src/continuedev/core/autopilot.py @@ -120,7 +120,7 @@ class Autopilot(ContinueBaseModel):          # If a parent step is deleted/cancelled, don't run this step          last_depth = self._step_depth          i = self.history.current_index -        while i >= 0 and self.history.timeline[i].depth > last_depth: +        while i >= 0 and self.history.timeline[i].depth == last_depth + 1:              if self.history.timeline[i].deleted:                  return None              last_depth = self.history.timeline[i].depth diff --git a/continuedev/src/continuedev/libs/llm/openai.py b/continuedev/src/continuedev/libs/llm/openai.py index 22c28b20..5d65eb22 100644 --- a/continuedev/src/continuedev/libs/llm/openai.py +++ b/continuedev/src/continuedev/libs/llm/openai.py @@ -83,14 +83,14 @@ class OpenAI(LLM):      def with_system_message(self, system_message: Union[str, None]):          return OpenAI(api_key=self.api_key, default_model=self.default_model, system_message=system_message) -    def stream_chat(self, prompt, with_history: List[ChatMessage] = [], **kwargs) -> Generator[Union[Any, List, Dict], None, None]: +    async def stream_chat(self, prompt, with_history: List[ChatMessage] = [], **kwargs) -> Generator[Union[Any, List, Dict], None, None]:          self.completion_count += 1          args = {"max_tokens": DEFAULT_MAX_TOKENS, "temperature": 0.5, "top_p": 1,                  "frequency_penalty": 0, "presence_penalty": 0} | kwargs          args["stream"] = True          args["model"] = "gpt-3.5-turbo" -        for chunk in openai.ChatCompletion.create( +        async for chunk in await openai.ChatCompletion.acreate(              messages=self.compile_chat_messages(with_history, prompt),              **args,          ): diff --git a/continuedev/src/continuedev/server/main.py b/continuedev/src/continuedev/server/main.py index b8bfe325..fb6ead02 100644 --- a/continuedev/src/continuedev/server/main.py +++ b/continuedev/src/continuedev/server/main.py @@ -40,10 +40,7 @@ args = parser.parse_args()  def run_server(): -    if os.path.exists("logging.yaml"): -        uvicorn.run(app, host="0.0.0.0", port=args.port, log_level="info") -    else: -        uvicorn.run(app, host="0.0.0.0", port=args.port, log_level="info") +    uvicorn.run(app, host="0.0.0.0", port=args.port)  if __name__ == "__main__": diff --git a/continuedev/src/continuedev/steps/chat.py b/continuedev/src/continuedev/steps/chat.py index 90514ad6..fd7457d9 100644 --- a/continuedev/src/continuedev/steps/chat.py +++ b/continuedev/src/continuedev/steps/chat.py @@ -11,6 +11,8 @@ class SimpleChatStep(Step):      async def run(self, sdk: ContinueSDK):          self.description = f"```{self.user_input}```\n\n" +        await sdk.update_ui() +          async for chunk in sdk.models.default.stream_chat(self.user_input, with_history=await sdk.get_chat_context()):              self.description += chunk              await sdk.update_ui() diff --git a/continuedev/src/continuedev/steps/react.py b/continuedev/src/continuedev/steps/react.py index 4d310fc8..cddb8b42 100644 --- a/continuedev/src/continuedev/steps/react.py +++ b/continuedev/src/continuedev/steps/react.py @@ -10,7 +10,8 @@ class NLDecisionStep(Step):      default_step: Union[Step, None] = None      steps: List[Tuple[Step, str]] -    hide: bool = True +    hide: bool = False +    name: str = "Deciding what to do next"      async def run(self, sdk: ContinueSDK):          step_descriptions = "\n".join([ @@ -36,4 +37,7 @@ class NLDecisionStep(Step):          step_to_run = step_to_run or self.default_step or self.steps[0] +        self.hide = True +        await sdk.update_ui() +          await sdk.run_step(step_to_run) | 
