You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"output.lang": "summarizing language uses English by default",
28
-
"openai.base_url": "what API base url to use.",
29
-
"openai.timeout": "request timeout",
30
-
"openai.max_tokens": "the maximum number of tokens to generate in the chat completion.",
31
-
"openai.temperature": "What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 will make it more focused and deterministic.",
32
-
"openai.provider": "service provider, only support 'openai' or 'azure'",
"openai.headers": "custom headers for openai request",
35
-
"openai.api_version": "openai api version",
36
-
"openai.top_p": "An alternative to sampling with temperature, called nucleus sampling, where the model considers the results of the tokens with top_p probability mass. So 0.1 means only the tokens comprising the top 10% probability mass are considered.",
37
-
"openai.frequency_penalty": "Number between 0.0 and 1.0 that penalizes new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim.",
38
-
"openai.presence_penalty": "Number between 0.0 and 1.0 that penalizes new tokens based on whether they appear in the text so far. Increases the model's likelihood to talk about new topics.",
39
-
"prompt.folder": "prompt template folder",
19
+
"git.diff_unified": "Generate diffs with <n> lines of context, default is 3",
20
+
"git.exclude_list": "Exclude file from git diff command",
21
+
"git.template_file": "Template file for commit message",
22
+
"git.template_string": "Template string for commit message",
"output.lang": "Summarizing language, defaults to English",
29
+
"openai.base_url": "API base URL to use",
30
+
"openai.timeout": "Request timeout",
31
+
"openai.max_tokens": "Maximum number of tokens to generate in the chat completion",
32
+
"openai.temperature": "Sampling temperature to use, between 0 and 2. Higher values like 0.8 make the output more random, while lower values like 0.2 make it more focused and deterministic",
33
+
"openai.provider": "Service provider, supports 'openai' or 'azure'",
"openai.headers": "Custom headers for OpenAI request",
36
+
"openai.api_version": "OpenAI API version",
37
+
"openai.top_p": "Nucleus sampling probability mass. For example, 0.1 means only the tokens comprising the top 10% probability mass are considered",
38
+
"openai.frequency_penalty": "Penalty for new tokens based on their existing frequency in the text so far. Decreases the model's likelihood to repeat the same line verbatim",
39
+
"openai.presence_penalty": "Penalty for new tokens based on whether they appear in the text so far. Increases the model's likelihood to talk about new topics",
40
+
"prompt.folder": "Prompt template folder",
40
41
}
41
42
42
43
// configListCmd represents the command to list the configuration values.
0 commit comments