summaryrefslogtreecommitdiff
path: root/continuedev/src
diff options
context:
space:
mode:
Diffstat (limited to 'continuedev/src')
-rw-r--r--continuedev/src/continuedev/core/autopilot.py2
-rw-r--r--continuedev/src/continuedev/libs/llm/openai.py4
-rw-r--r--continuedev/src/continuedev/server/main.py5
-rw-r--r--continuedev/src/continuedev/steps/chat.py2
-rw-r--r--continuedev/src/continuedev/steps/react.py6
5 files changed, 11 insertions, 8 deletions
diff --git a/continuedev/src/continuedev/core/autopilot.py b/continuedev/src/continuedev/core/autopilot.py
index 73f46a37..ee249c0b 100644
--- a/continuedev/src/continuedev/core/autopilot.py
+++ b/continuedev/src/continuedev/core/autopilot.py
@@ -120,7 +120,7 @@ class Autopilot(ContinueBaseModel):
# If a parent step is deleted/cancelled, don't run this step
last_depth = self._step_depth
i = self.history.current_index
- while i >= 0 and self.history.timeline[i].depth > last_depth:
+ while i >= 0 and self.history.timeline[i].depth == last_depth + 1:
if self.history.timeline[i].deleted:
return None
last_depth = self.history.timeline[i].depth
diff --git a/continuedev/src/continuedev/libs/llm/openai.py b/continuedev/src/continuedev/libs/llm/openai.py
index 22c28b20..5d65eb22 100644
--- a/continuedev/src/continuedev/libs/llm/openai.py
+++ b/continuedev/src/continuedev/libs/llm/openai.py
@@ -83,14 +83,14 @@ class OpenAI(LLM):
def with_system_message(self, system_message: Union[str, None]):
return OpenAI(api_key=self.api_key, default_model=self.default_model, system_message=system_message)
- def stream_chat(self, prompt, with_history: List[ChatMessage] = [], **kwargs) -> Generator[Union[Any, List, Dict], None, None]:
+ async def stream_chat(self, prompt, with_history: List[ChatMessage] = [], **kwargs) -> Generator[Union[Any, List, Dict], None, None]:
self.completion_count += 1
args = {"max_tokens": DEFAULT_MAX_TOKENS, "temperature": 0.5, "top_p": 1,
"frequency_penalty": 0, "presence_penalty": 0} | kwargs
args["stream"] = True
args["model"] = "gpt-3.5-turbo"
- for chunk in openai.ChatCompletion.create(
+ async for chunk in await openai.ChatCompletion.acreate(
messages=self.compile_chat_messages(with_history, prompt),
**args,
):
diff --git a/continuedev/src/continuedev/server/main.py b/continuedev/src/continuedev/server/main.py
index b8bfe325..fb6ead02 100644
--- a/continuedev/src/continuedev/server/main.py
+++ b/continuedev/src/continuedev/server/main.py
@@ -40,10 +40,7 @@ args = parser.parse_args()
def run_server():
- if os.path.exists("logging.yaml"):
- uvicorn.run(app, host="0.0.0.0", port=args.port, log_level="info")
- else:
- uvicorn.run(app, host="0.0.0.0", port=args.port, log_level="info")
+ uvicorn.run(app, host="0.0.0.0", port=args.port)
if __name__ == "__main__":
diff --git a/continuedev/src/continuedev/steps/chat.py b/continuedev/src/continuedev/steps/chat.py
index 90514ad6..fd7457d9 100644
--- a/continuedev/src/continuedev/steps/chat.py
+++ b/continuedev/src/continuedev/steps/chat.py
@@ -11,6 +11,8 @@ class SimpleChatStep(Step):
async def run(self, sdk: ContinueSDK):
self.description = f"```{self.user_input}```\n\n"
+ await sdk.update_ui()
+
async for chunk in sdk.models.default.stream_chat(self.user_input, with_history=await sdk.get_chat_context()):
self.description += chunk
await sdk.update_ui()
diff --git a/continuedev/src/continuedev/steps/react.py b/continuedev/src/continuedev/steps/react.py
index 4d310fc8..cddb8b42 100644
--- a/continuedev/src/continuedev/steps/react.py
+++ b/continuedev/src/continuedev/steps/react.py
@@ -10,7 +10,8 @@ class NLDecisionStep(Step):
default_step: Union[Step, None] = None
steps: List[Tuple[Step, str]]
- hide: bool = True
+ hide: bool = False
+ name: str = "Deciding what to do next"
async def run(self, sdk: ContinueSDK):
step_descriptions = "\n".join([
@@ -36,4 +37,7 @@ class NLDecisionStep(Step):
step_to_run = step_to_run or self.default_step or self.steps[0]
+ self.hide = True
+ await sdk.update_ui()
+
await sdk.run_step(step_to_run)