From 4657f30167d18c42aa66cdd841965c9661669d3f Mon Sep 17 00:00:00 2001 From: Nate Sesti Date: Sun, 2 Jul 2023 15:47:14 -0700 Subject: no more gpt-3.5-turbo --- continuedev/src/continuedev/steps/chat.py | 2 +- continuedev/src/continuedev/steps/core/core.py | 2 +- extension/react-app/src/components/TextDialog.tsx | 3 ++- extension/react-app/src/tabs/gui.tsx | 22 +++++++++++++++++----- 4 files changed, 21 insertions(+), 8 deletions(-) diff --git a/continuedev/src/continuedev/steps/chat.py b/continuedev/src/continuedev/steps/chat.py index 9bb75ab4..2efef37d 100644 --- a/continuedev/src/continuedev/steps/chat.py +++ b/continuedev/src/continuedev/steps/chat.py @@ -36,7 +36,7 @@ class SimpleChatStep(Step): summary=self.user_input )) - async for chunk in sdk.models.default.stream_chat(messages): + async for chunk in sdk.models.gpt4.stream_chat(messages): if sdk.current_step_was_deleted(): return diff --git a/continuedev/src/continuedev/steps/core/core.py b/continuedev/src/continuedev/steps/core/core.py index a84263cc..345d99bc 100644 --- a/continuedev/src/continuedev/steps/core/core.py +++ b/continuedev/src/continuedev/steps/core/core.py @@ -172,7 +172,7 @@ class DefaultModelEditCodeStep(Step): # We don't know here all of the functions being passed in. # We care because if this prompt itself goes over the limit, then the entire message will have to be cut from the completion. # Overflow won't happen, but prune_chat_messages in count_tokens.py will cut out this whole thing, instead of us cutting out only as many lines as we need. - model_to_use = sdk.models.default + model_to_use = sdk.models.gpt4 BUFFER_FOR_FUNCTIONS = 400 total_tokens = model_to_use.count_tokens( diff --git a/extension/react-app/src/components/TextDialog.tsx b/extension/react-app/src/components/TextDialog.tsx index 2632e572..a564f884 100644 --- a/extension/react-app/src/components/TextDialog.tsx +++ b/extension/react-app/src/components/TextDialog.tsx @@ -52,6 +52,7 @@ const TextDialog = (props: { showDialog: boolean; onEnter: (text: string) => void; onClose: () => void; + message?: string; }) => { const [text, setText] = useState(""); const textAreaRef = React.createRef(); @@ -75,7 +76,7 @@ const TextDialog = (props: { }} > -

Thanks for your feedback. We'll get back to you soon!

+

{props.message || ""}