@@ -178,6 +178,7 @@ export const GOOGLE_MODELS = [
178178 "gemini-2.0-flash-lite-8k" ,
179179 "gemini-3-flash-preview-16k" , // Preview model, context limited to 16k
180180 "gemini-3-pro-preview-8k" , // Preview model, context limited to 8k
181+ "gemini-3.1-pro-preview-8k" , // Preview model, context limited to 8k
181182] as const ;
182183export type GoogleModel = ( typeof GOOGLE_MODELS ) [ number ] ;
183184export function isGoogleModel ( model : unknown ) : model is GoogleModel {
@@ -197,6 +198,7 @@ const CANONICAL_GOOGLE_MODELS_THINKING = [
197198 "gemini-2.5-pro" ,
198199 "gemini-3-flash-preview" ,
199200 "gemini-3-pro-preview" ,
201+ "gemini-3.1-pro-preview" ,
200202] as const ;
201203
202204export type CanonicalGoogleModel = ( typeof CANONICAL_GOOGLE_MODELS ) [ number ] ;
@@ -220,6 +222,7 @@ export const GOOGLE_MODEL_TO_ID: Partial<{
220222 "gemini-2.5-pro-8k" : "gemini-2.5-pro" ,
221223 "gemini-3-flash-preview-16k" : "gemini-3-flash-preview" ,
222224 "gemini-3-pro-preview-8k" : "gemini-3-pro-preview" ,
225+ "gemini-3.1-pro-preview-8k" : "gemini-3.1-pro-preview" ,
223226} as const ;
224227
225228/**
@@ -359,6 +362,7 @@ export const USER_SELECTABLE_LLMS_BY_VENDOR: {
359362 m === "gpt-5-mini-8k" ,
360363 ) ,
361364 google : [
365+ "gemini-3.1-pro-preview-8k" ,
362366 "gemini-3-flash-preview-16k" ,
363367 "gemini-3-pro-preview-8k" ,
364368 "gemini-2.5-flash-8k" ,
@@ -875,6 +879,7 @@ export const LLM_USERNAMES: LLM2String = {
875879 "gemini-2.5-flash-8k" : "Gemini 2.5 Flash" ,
876880 "gemini-2.5-pro-8k" : "Gemini 2.5 Pro" ,
877881 "gemini-3-pro-preview-8k" : "Gemini 3 Pro" ,
882+ "gemini-3.1-pro-preview-8k" : "Gemini 3.1 Pro" ,
878883 "mistral-small-latest" : "Mistral AI Small" ,
879884 "mistral-medium-latest" : "Mistral AI Medium" ,
880885 "mistral-large-latest" : "Mistral AI Large" ,
@@ -967,6 +972,8 @@ export const LLM_DESCR: LLM2String = {
967972 "Google's Gemini 2.5 Pro Generative AI model (8k token context)" ,
968973 "gemini-3-pro-preview-8k" :
969974 "Google's Gemini 3 Pro Generative AI model (8k token context)" ,
975+ "gemini-3.1-pro-preview-8k" :
976+ "Google's Gemini 3.1 Pro model with enhanced reasoning (8k token context)" ,
970977 "mistral-small-latest" :
971978 "Small general purpose tasks, text classification, customer service. (Mistral AI, 4k token context)" ,
972979 "mistral-medium-latest" :
@@ -1299,6 +1306,12 @@ export const LLM_COST: { [name in LanguageModelCore]: Cost } = {
12991306 max_tokens : 8_000 ,
13001307 free : false ,
13011308 } ,
1309+ "gemini-3.1-pro-preview-8k" : {
1310+ prompt_tokens : usd1Mtokens ( 2 ) ,
1311+ completion_tokens : usd1Mtokens ( 12 ) ,
1312+ max_tokens : 8_000 ,
1313+ free : false ,
1314+ } ,
13021315 // https://mistral.ai/technology/
13031316 "mistral-small-latest" : {
13041317 prompt_tokens : usd1Mtokens ( 0.2 ) ,
0 commit comments