From 90590ab4e06fbc3fa721f73a4a922136946a756f Mon Sep 17 00:00:00 2001 From: Nate Sesti Date: Sat, 2 Sep 2023 11:52:16 -0700 Subject: fix: :bug: llamacpp fix indexing max_tokens --- continuedev/src/continuedev/libs/llm/llamacpp.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'continuedev') diff --git a/continuedev/src/continuedev/libs/llm/llamacpp.py b/continuedev/src/continuedev/libs/llm/llamacpp.py index 84eb85fd..72c766c0 100644 --- a/continuedev/src/continuedev/libs/llm/llamacpp.py +++ b/continuedev/src/continuedev/libs/llm/llamacpp.py @@ -71,7 +71,7 @@ class LlamaCpp(LLM): self.model, with_history, self.context_length, - args["max_tokens"], + args["n_predict"] if "n_predict" in args else 1024, prompt, functions=args.get("functions", None), system_message=self.system_message, -- cgit v1.2.3-70-g09d2