diff options
Diffstat (limited to 'continuedev')
| -rw-r--r-- | continuedev/src/continuedev/libs/llm/__init__.py | 10 | 
1 files changed, 7 insertions, 3 deletions
diff --git a/continuedev/src/continuedev/libs/llm/__init__.py b/continuedev/src/continuedev/libs/llm/__init__.py index 28f614c7..e6a90ef7 100644 --- a/continuedev/src/continuedev/libs/llm/__init__.py +++ b/continuedev/src/continuedev/libs/llm/__init__.py @@ -71,6 +71,10 @@ class LLM(ContinueBaseModel):          ..., description="The name of the model to be used (e.g. gpt-4, codellama)"      ) +    max_tokens: int = Field( +        DEFAULT_MAX_TOKENS, description="The maximum number of tokens to generate." +    ) +      stop_tokens: Optional[List[str]] = Field(          None, description="Tokens that will stop the completion."      ) @@ -237,7 +241,7 @@ class LLM(ContinueBaseModel):              presence_penalty=presence_penalty,              frequency_penalty=frequency_penalty,              stop=stop or self.stop_tokens, -            max_tokens=max_tokens, +            max_tokens=max_tokens or self.max_tokens,              functions=functions,          ) @@ -288,7 +292,7 @@ class LLM(ContinueBaseModel):              presence_penalty=presence_penalty,              frequency_penalty=frequency_penalty,              stop=stop or self.stop_tokens, -            max_tokens=max_tokens, +            max_tokens=max_tokens or self.max_tokens,              functions=functions,          ) @@ -337,7 +341,7 @@ class LLM(ContinueBaseModel):              presence_penalty=presence_penalty,              frequency_penalty=frequency_penalty,              stop=stop or self.stop_tokens, -            max_tokens=max_tokens, +            max_tokens=max_tokens or self.max_tokens,              functions=functions,          )  | 
