From b9bdf5894c1c68b60d1919ae07b0f5909b00dec2 Mon Sep 17 00:00:00 2001 From: Nate Sesti Date: Sat, 2 Sep 2023 11:27:30 -0700 Subject: fix: :bug: number of bug fixes --- continuedev/src/continuedev/libs/llm/hf_tgi.py | 2 +- continuedev/src/continuedev/libs/llm/llamacpp.py | 8 ++++---- continuedev/src/continuedev/plugins/context_providers/file.py | 2 +- 3 files changed, 6 insertions(+), 6 deletions(-) (limited to 'continuedev') diff --git a/continuedev/src/continuedev/libs/llm/hf_tgi.py b/continuedev/src/continuedev/libs/llm/hf_tgi.py index f106f83f..5c7e0239 100644 --- a/continuedev/src/continuedev/libs/llm/hf_tgi.py +++ b/continuedev/src/continuedev/libs/llm/hf_tgi.py @@ -52,7 +52,7 @@ class HuggingFaceTGI(LLM): ) as client_session: async with client_session.post( f"{self.server_url}", - json={"inputs": prompt, **self._transform_args(args)}, + json={"inputs": prompt, **args}, ) as resp: async for line in resp.content.iter_any(): if line: diff --git a/continuedev/src/continuedev/libs/llm/llamacpp.py b/continuedev/src/continuedev/libs/llm/llamacpp.py index 7940c4c9..84eb85fd 100644 --- a/continuedev/src/continuedev/libs/llm/llamacpp.py +++ b/continuedev/src/continuedev/libs/llm/llamacpp.py @@ -87,7 +87,7 @@ class LlamaCpp(LLM): f"{self.server_url}/completion", json={ "prompt": prompt, - **self._transform_args(args), + **args, }, headers={"Content-Type": "application/json"}, ) as resp: @@ -107,7 +107,7 @@ class LlamaCpp(LLM): self.model, messages, self.context_length, - args["max_tokens"], + args["n_predict"] if "n_predict" in args else 1024, None, functions=args.get("functions", None), system_message=self.system_message, @@ -123,7 +123,7 @@ class LlamaCpp(LLM): ) as client_session: async with client_session.post( f"{self.server_url}/completion", - json={"prompt": prompt, **self._transform_args(args)}, + json={"prompt": prompt, **args}, headers=headers, ) as resp: async for line in resp.content: @@ -170,7 +170,7 @@ class LlamaCpp(LLM): ) as client_session: async with client_session.post( f"{self.server_url}/completion", - json={"prompt": prompt, **self._transform_args(args)}, + json={"prompt": prompt, **args}, headers={"Content-Type": "application/json"}, ) as resp: json_resp = await resp.json() diff --git a/continuedev/src/continuedev/plugins/context_providers/file.py b/continuedev/src/continuedev/plugins/context_providers/file.py index 859088b8..a300a4a9 100644 --- a/continuedev/src/continuedev/plugins/context_providers/file.py +++ b/continuedev/src/continuedev/plugins/context_providers/file.py @@ -123,7 +123,7 @@ class FileContextProvider(ContextProvider): ) async def provide_context_items(self, workspace_dir: str) -> List[ContextItem]: - contents = await self.sdk.ide.listDirectoryContents(workspace_dir, True) + contents = await self.sdk.ide.listDirectoryContents(workspace_dir, False) if contents is None: return [] -- cgit v1.2.3-70-g09d2