Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
48 changes: 42 additions & 6 deletions packages/types/src/providers/vscode-llm.ts
Original file line number Diff line number Diff line change
Expand Up @@ -101,6 +101,18 @@ export const vscodeLlmModels = {
supportsToolCalling: true,
maxInputTokens: 81638,
},
"claude-4-sonnet": {
contextWindow: 128000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 0,
outputPrice: 0,
family: "claude-sonnet-4",
version: "claude-sonnet-4",
name: "Claude Sonnet 4",
supportsToolCalling: true,
maxInputTokens: 111836,
},
"gemini-2.0-flash-001": {
contextWindow: 127827,
supportsImages: true,
Expand All @@ -114,7 +126,7 @@ export const vscodeLlmModels = {
maxInputTokens: 127827,
},
"gemini-2.5-pro": {
contextWindow: 63830,
contextWindow: 128000,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you add a comment explaining why the contextWindow was increased from 63830 to 128000? This is a significant change that might affect existing users of this model.

supportsImages: true,
supportsPromptCache: false,
inputPrice: 0,
Expand All @@ -123,10 +135,10 @@ export const vscodeLlmModels = {
version: "gemini-2.5-pro-preview-03-25",
name: "Gemini 2.5 Pro (Preview)",
supportsToolCalling: true,
maxInputTokens: 63830,
maxInputTokens: 108637,
},
"o4-mini": {
contextWindow: 111446,
contextWindow: 128000,
supportsImages: false,
supportsPromptCache: false,
inputPrice: 0,
Expand All @@ -135,10 +147,10 @@ export const vscodeLlmModels = {
version: "o4-mini-2025-04-16",
name: "o4-mini (Preview)",
supportsToolCalling: true,
maxInputTokens: 111446,
maxInputTokens: 111452,
},
"gpt-4.1": {
contextWindow: 111446,
contextWindow: 128000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 0,
Expand All @@ -147,7 +159,31 @@ export const vscodeLlmModels = {
version: "gpt-4.1-2025-04-14",
name: "GPT-4.1 (Preview)",
supportsToolCalling: true,
maxInputTokens: 111446,
maxInputTokens: 111452,
},
"gpt-5-mini": {
contextWindow: 128000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 0,
outputPrice: 0,
family: "gpt-5-mini",
version: "gpt-5-mini",
name: "GPT-5 mini (Preview)",
supportsToolCalling: true,
maxInputTokens: 108637,
},
"gpt-5": {
contextWindow: 128000,
supportsImages: true,
supportsPromptCache: false,
inputPrice: 0,
outputPrice: 0,
family: "gpt-5",
version: "gpt-5",
name: "GPT-5 (Preview)",
supportsToolCalling: true,
maxInputTokens: 108637,
},
} as const satisfies Record<
string,
Expand Down
Loading