@@ -160,17 +160,12 @@ def check_available_online(
160160 "Fairseq2LlamaForCausalLM" : _HfExamplesInfo ("mgleize/fairseq2-dummy-Llama-3.2-1B" ), # noqa: E501
161161 "FalconForCausalLM" : _HfExamplesInfo ("tiiuae/falcon-7b" ),
162162 "FalconH1ForCausalLM" :_HfExamplesInfo ("tiiuae/Falcon-H1-1.5B-Instruct" ,
163- is_available_online = False ,
164- min_transformers_version = "4.52.2" ),
163+ min_transformers_version = "4.53" ),
165164 "GemmaForCausalLM" : _HfExamplesInfo ("google/gemma-1.1-2b-it" ),
166165 "Gemma2ForCausalLM" : _HfExamplesInfo ("google/gemma-2-9b" ),
167166 "Gemma3ForCausalLM" : _HfExamplesInfo ("google/gemma-3-1b-it" ),
168167 "GlmForCausalLM" : _HfExamplesInfo ("THUDM/glm-4-9b-chat-hf" ),
169- "Glm4ForCausalLM" : _HfExamplesInfo (
170- "THUDM/GLM-4-32B-0414" ,
171- is_available_online = False ,
172- min_transformers_version = "4.52.dev0"
173- ),
168+ "Glm4ForCausalLM" : _HfExamplesInfo ("THUDM/GLM-4-9B-0414" ),
174169 "GPT2LMHeadModel" : _HfExamplesInfo ("openai-community/gpt2" ,
175170 {"alias" : "gpt2" }),
176171 "GPTBigCodeForCausalLM" : _HfExamplesInfo ("bigcode/starcoder" ,
@@ -181,8 +176,7 @@ def check_available_online(
181176 {"1b" : "EleutherAI/pythia-1.4b" }),
182177 "GraniteForCausalLM" : _HfExamplesInfo ("ibm/PowerLM-3b" ),
183178 "GraniteMoeForCausalLM" : _HfExamplesInfo ("ibm/PowerMoE-3b" ),
184- "GraniteMoeHybridForCausalLM" : _HfExamplesInfo ("ibm-granite/granite-4.0-tiny-preview" , # noqa: E501
185- min_transformers_version = "4.52.0" ), # noqa: E501
179+ "GraniteMoeHybridForCausalLM" : _HfExamplesInfo ("ibm-granite/granite-4.0-tiny-preview" ), # noqa: E501
186180 "GraniteMoeSharedForCausalLM" : _HfExamplesInfo ("ibm-research/moe-7b-1b-active-shared-experts" ), # noqa: E501
187181 "Grok1ModelForCausalLM" : _HfExamplesInfo ("hpcai-tech/grok-1" ,
188182 trust_remote_code = True ),
@@ -203,8 +197,7 @@ def check_available_online(
203197 "LLaMAForCausalLM" : _HfExamplesInfo ("decapoda-research/llama-7b-hf" ,
204198 is_available_online = False ),
205199 "MambaForCausalLM" : _HfExamplesInfo ("state-spaces/mamba-130m-hf" ),
206- "Mamba2ForCausalLM" : _HfExamplesInfo ("mistralai/Mamba-Codestral-7B-v0.1" ,
207- is_available_online = False ),
200+ "Mamba2ForCausalLM" : _HfExamplesInfo ("mistralai/Mamba-Codestral-7B-v0.1" ),
208201 "FalconMambaForCausalLM" : _HfExamplesInfo ("tiiuae/falcon-mamba-7b-instruct" ), # noqa: E501
209202 "MiniCPMForCausalLM" : _HfExamplesInfo ("openbmb/MiniCPM-2B-sft-bf16" ,
210203 trust_remote_code = True ),
@@ -243,10 +236,9 @@ def check_available_online(
243236 "Qwen2MoeForCausalLM" : _HfExamplesInfo ("Qwen/Qwen1.5-MoE-A2.7B-Chat" ),
244237 "Qwen3ForCausalLM" : _HfExamplesInfo ("Qwen/Qwen3-8B" ),
245238 "Qwen3MoeForCausalLM" : _HfExamplesInfo ("Qwen/Qwen3-30B-A3B" ),
246- "RWForCausalLM" : _HfExamplesInfo ("tiiuae/falcon-40b" ,
247- is_available_online = False ),
239+ "RWForCausalLM" : _HfExamplesInfo ("tiiuae/falcon-40b" ),
248240 "StableLMEpochForCausalLM" : _HfExamplesInfo ("stabilityai/stablelm-zephyr-3b" , # noqa: E501
249- is_available_online = False ),
241+ v0_only = True ),
250242 "StableLmForCausalLM" : _HfExamplesInfo ("stabilityai/stablelm-3b-4e1t" ,
251243 v0_only = True ),
252244 "Starcoder2ForCausalLM" : _HfExamplesInfo ("bigcode/starcoder2-3b" ),
@@ -256,7 +248,7 @@ def check_available_online(
256248 "TeleFLMForCausalLM" : _HfExamplesInfo ("CofeAI/FLM-2-52B-Instruct-2407" ,
257249 trust_remote_code = True ),
258250 "XverseForCausalLM" : _HfExamplesInfo ("xverse/XVERSE-7B-Chat" ,
259- is_available_online = False ,
251+ tokenizer = "meta-llama/Llama-2-7b" ,
260252 trust_remote_code = True ),
261253 "Zamba2ForCausalLM" : _HfExamplesInfo ("Zyphra/Zamba2-7B-instruct" ),
262254 "MiMoForCausalLM" : _HfExamplesInfo ("XiaomiMiMo/MiMo-7B-RL" ,
@@ -275,8 +267,7 @@ def check_available_online(
275267 trust_remote_code = True ),
276268 "GteNewModel" : _HfExamplesInfo ("Alibaba-NLP/gte-base-en-v1.5" ,
277269 trust_remote_code = True ,
278- hf_overrides = {"architectures" :
279- ["GteNewModel" ]}),
270+ hf_overrides = {"architectures" : ["GteNewModel" ]}), # noqa: E501
280271 "InternLM2ForRewardModel" : _HfExamplesInfo ("internlm/internlm2-1_8b-reward" ,
281272 trust_remote_code = True ),
282273 "JambaForSequenceClassification" : _HfExamplesInfo ("ai21labs/Jamba-tiny-reward-dev" ), # noqa: E501
@@ -298,10 +289,8 @@ def check_available_online(
298289 "Phi3VForCausalLM" : _HfExamplesInfo ("TIGER-Lab/VLM2Vec-Full" ,
299290 trust_remote_code = True ),
300291 "Qwen2VLForConditionalGeneration" : _HfExamplesInfo ("MrLight/dse-qwen2-2b-mrl-v1" ), # noqa: E501
301- # The model on Huggingface is currently being updated,
302- # hence I temporarily mark it as not available online
303- "PrithviGeoSpatialMAE" : _HfExamplesInfo ("ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11" , # noqa: E501
304- is_available_online = False ),
292+ "PrithviGeoSpatialMAE" : _HfExamplesInfo ("ibm-nasa-geospatial/Prithvi-EO-2.0-300M-TL-Sen1Floods11" , # noqa: E501
293+ is_available_online = False ), # noqa: E501
305294}
306295
307296_CROSS_ENCODER_EXAMPLE_MODELS = {
@@ -327,8 +316,7 @@ def check_available_online(
327316 hf_overrides = {"architectures" : ["DeepseekVLV2ForCausalLM" ]}), # noqa: E501
328317 "FuyuForCausalLM" : _HfExamplesInfo ("adept/fuyu-8b" ),
329318 "Gemma3ForConditionalGeneration" : _HfExamplesInfo ("google/gemma-3-4b-it" ),
330- "GraniteSpeechForConditionalGeneration" : _HfExamplesInfo ("ibm-granite/granite-speech-3.3-8b" , # noqa: E501
331- min_transformers_version = "4.52.0" ), # noqa: E501
319+ "GraniteSpeechForConditionalGeneration" : _HfExamplesInfo ("ibm-granite/granite-speech-3.3-2b" ), # noqa: E501
332320 "GLM4VForCausalLM" : _HfExamplesInfo ("THUDM/glm-4v-9b" ,
333321 trust_remote_code = True ,
334322 hf_overrides = {"architectures" : ["GLM4VForCausalLM" ]}), # noqa: E501
@@ -347,7 +335,6 @@ def check_available_online(
347335 trust_remote_code = True ,
348336 v0_only = True ),
349337 "Llama4ForConditionalGeneration" : _HfExamplesInfo ("meta-llama/Llama-4-Scout-17B-16E-Instruct" , # noqa: E501
350- min_transformers_version = "4.51" ,
351338 max_model_len = 10240 ),
352339 "LlavaForConditionalGeneration" : _HfExamplesInfo ("llava-hf/llava-1.5-7b-hf" ,
353340 extras = {"mistral" : "mistral-community/pixtral-12b" , # noqa: E501
@@ -360,8 +347,6 @@ def check_available_online(
360347 transformers_version_reason = "HF model is not compatible." , # noqa: E501
361348 hf_overrides = {"architectures" : ["MantisForConditionalGeneration" ]}), # noqa: E501
362349 "MiniCPMO" : _HfExamplesInfo ("openbmb/MiniCPM-o-2_6" ,
363- max_transformers_version = "4.48" ,
364- transformers_version_reason = "Use of deprecated imports which have been removed." , # noqa: E501
365350 trust_remote_code = True ),
366351 "MiniCPMV" : _HfExamplesInfo ("openbmb/MiniCPM-Llama3-V-2_5" ,
367352 extras = {"2.6" : "openbmb/MiniCPM-V-2_6" }, # noqa: E501
@@ -399,10 +384,8 @@ def check_available_online(
399384 "Qwen2AudioForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2-Audio-7B-Instruct" ), # noqa: E501
400385 "Qwen2VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2-VL-2B-Instruct" ), # noqa: E501
401386 "Qwen2_5_VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-VL-3B-Instruct" ), # noqa: E501
402- "Qwen2_5OmniModel" : _HfExamplesInfo ("Qwen/Qwen2.5-Omni-3B" ,
403- min_transformers_version = "4.52" ),
404- "Qwen2_5OmniForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-Omni-7B-AWQ" , # noqa: E501
405- min_transformers_version = "4.52" ),
387+ "Qwen2_5OmniModel" : _HfExamplesInfo ("Qwen/Qwen2.5-Omni-3B" ),
388+ "Qwen2_5OmniForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-Omni-7B-AWQ" ), # noqa: E501
406389 "SkyworkR1VChatModel" : _HfExamplesInfo ("Skywork/Skywork-R1V-38B" ),
407390 "SmolVLMForConditionalGeneration" : _HfExamplesInfo ("HuggingFaceTB/SmolVLM2-2.2B-Instruct" ), # noqa: E501
408391 "UltravoxModel" : _HfExamplesInfo ("fixie-ai/ultravox-v0_5-llama-3_2-1b" , # noqa: E501
@@ -413,8 +396,8 @@ def check_available_online(
413396 # Florence-2 uses BartFastTokenizer which can't be loaded from AutoTokenizer
414397 # Therefore, we borrow the BartTokenizer from the original Bart model
415398 "Florence2ForConditionalGeneration" : _HfExamplesInfo ("microsoft/Florence-2-base" , # noqa: E501
416- tokenizer = "Isotr0py/Florence-2-tokenizer" ,
417- trust_remote_code = True , ), # noqa: E501
399+ tokenizer = "Isotr0py/Florence-2-tokenizer" , # noqa: E501
400+ trust_remote_code = True ), # noqa: E501
418401 "MllamaForConditionalGeneration" : _HfExamplesInfo ("meta-llama/Llama-3.2-11B-Vision-Instruct" ), # noqa: E501
419402 "WhisperForConditionalGeneration" : _HfExamplesInfo ("openai/whisper-large-v3" ), # noqa: E501
420403}
0 commit comments