summaryrefslogtreecommitdiff
path: root/continuedev/src/continuedev/libs/util
diff options
context:
space:
mode:
authorNate Sesti <sestinj@gmail.com>2023-06-26 11:09:32 -0700
committerNate Sesti <sestinj@gmail.com>2023-06-26 11:09:32 -0700
commit281f5aa23777cedf794bc51c75eb860aeed758c2 (patch)
tree1d3e0ac16d08ef213fc975b1e45e3461e0b04c84 /continuedev/src/continuedev/libs/util
parentd4ac2379e3958ef34987d366dce26d5f6d472375 (diff)
parente21cde260ad518d83a9811d11362d91a47f0b0c5 (diff)
downloadsncontinue-281f5aa23777cedf794bc51c75eb860aeed758c2.tar.gz
sncontinue-281f5aa23777cedf794bc51c75eb860aeed758c2.tar.bz2
sncontinue-281f5aa23777cedf794bc51c75eb860aeed758c2.zip
Merge branch 'main' of https://github.com/continuedev/continue
Diffstat (limited to 'continuedev/src/continuedev/libs/util')
-rw-r--r--continuedev/src/continuedev/libs/util/count_tokens.py4
1 files changed, 2 insertions, 2 deletions
diff --git a/continuedev/src/continuedev/libs/util/count_tokens.py b/continuedev/src/continuedev/libs/util/count_tokens.py
index addafcff..154af5e1 100644
--- a/continuedev/src/continuedev/libs/util/count_tokens.py
+++ b/continuedev/src/continuedev/libs/util/count_tokens.py
@@ -21,7 +21,7 @@ def encoding_for_model(model: str):
return tiktoken.encoding_for_model(aliases.get(model, model))
-def count_tokens(model: str, text: str | None):
+def count_tokens(model: str, text: Union[str, None]):
if text is None:
return 0
encoding = encoding_for_model(model)
@@ -74,7 +74,7 @@ def prune_chat_history(model: str, chat_history: List[ChatMessage], max_tokens:
return chat_history
-def compile_chat_messages(model: str, msgs: List[ChatMessage], prompt: str | None = None, with_functions: bool = False, system_message: Union[str, None] = None) -> List[Dict]:
+def compile_chat_messages(model: str, msgs: List[ChatMessage], prompt: Union[str, None] = None, with_functions: bool = False, system_message: Union[str, None] = None) -> List[Dict]:
prompt_tokens = count_tokens(model, prompt)
msgs = prune_chat_history(model,
msgs, MAX_TOKENS_FOR_MODEL[model], prompt_tokens + 1000 + count_tokens(model, system_message))