From 95363a5b52f3bf73531ac76b00178fa79ca97661 Mon Sep 17 00:00:00 2001 From: Nate Sesti <33237525+sestinj@users.noreply.github.com> Date: Thu, 28 Sep 2023 01:02:52 -0700 Subject: Past input (#513) * feat: :construction: use ComboBox in place of UserInputContainer * feat: :construction: adding context to previous inputs steps * feat: :sparkles: preview context items on click * feat: :construction: more work on context items ui * style: :construction: working out the details of ctx item buttons * feat: :sparkles: getting the final details * fix: :bug: fix height of ctx items bar * fix: :bug: last couple of details * fix: :bug: pass model param through to hf inference api * fix: :loud_sound: better logging for timeout * feat: :sparkles: option to set the meilisearch url * fix: :bug: fix height of past inputs --- extension/schema/ContinueConfig.d.ts | 13 ++++++++----- 1 file changed, 8 insertions(+), 5 deletions(-) (limited to 'extension/schema/ContinueConfig.d.ts') diff --git a/extension/schema/ContinueConfig.d.ts b/extension/schema/ContinueConfig.d.ts index 92f6e047..64aa5c02 100644 --- a/extension/schema/ContinueConfig.d.ts +++ b/extension/schema/ContinueConfig.d.ts @@ -71,11 +71,15 @@ export type VerifySsl = boolean; * Path to a custom CA bundle to use when making the HTTP request */ export type CaBundlePath = string; +/** + * Proxy URL to use when making the HTTP request + */ +export type Proxy = string; /** * The API key for the LLM provider. */ export type ApiKey = string; -export type Unused = LLM[]; +export type Saved = LLM[]; /** * The temperature parameter for sampling from the LLM. Higher temperatures will result in more random output, while lower temperatures will result in more predictable output. This value ranges from 0 to 1. */ @@ -205,12 +209,10 @@ export interface FunctionCall { */ export interface Models1 { default: LLM; - small?: LLM; - medium?: LLM; - large?: LLM; + summarize?: LLM; edit?: LLM; chat?: LLM; - unused?: Unused; + saved?: Saved; sdk?: ContinueSDK; [k: string]: unknown; } @@ -224,6 +226,7 @@ export interface LLM { timeout?: Timeout; verify_ssl?: VerifySsl; ca_bundle_path?: CaBundlePath; + proxy?: Proxy; prompt_templates?: PromptTemplates; api_key?: ApiKey; [k: string]: unknown; -- cgit v1.2.3-70-g09d2