@@ -121,10 +121,11 @@ def check_available_online(
121121 "BaichuanForCausalLM" : _HfExamplesInfo ("baichuan-inc/Baichuan2-7B-chat" ,
122122 trust_remote_code = True ),
123123 "BambaForCausalLM" : _HfExamplesInfo ("ibm-ai-platform/Bamba-9B" ),
124- "BloomForCausalLM" : _HfExamplesInfo ("bigscience/bloomz-1b1" ),
124+ "BloomForCausalLM" : _HfExamplesInfo ("bigscience/bloom-560m" ,
125+ {"1b" : "bigscience/bloomz-1b1" }),
125126 "ChatGLMModel" : _HfExamplesInfo ("THUDM/chatglm3-6b" ,
126127 trust_remote_code = True ,
127- max_transformers_version = "4.51.1 " ),
128+ max_transformers_version = "4.48 " ),
128129 "ChatGLMForConditionalGeneration" : _HfExamplesInfo ("thu-coai/ShieldLM-6B-chatglm3" , # noqa: E501
129130 trust_remote_code = True ),
130131 "CohereForCausalLM" : _HfExamplesInfo ("CohereForAI/c4ai-command-r-v01" ,
@@ -142,24 +143,26 @@ def check_available_online(
142143 "ExaoneForCausalLM" : _HfExamplesInfo ("LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct" ), # noqa: E501
143144 "Fairseq2LlamaForCausalLM" : _HfExamplesInfo ("mgleize/fairseq2-dummy-Llama-3.2-1B" ), # noqa: E501
144145 "FalconForCausalLM" : _HfExamplesInfo ("tiiuae/falcon-7b" ),
145- "GemmaForCausalLM" : _HfExamplesInfo ("google/gemma-2b " ),
146+ "GemmaForCausalLM" : _HfExamplesInfo ("google/gemma-1.1-2b-it " ),
146147 "Gemma2ForCausalLM" : _HfExamplesInfo ("google/gemma-2-9b" ),
147- "Gemma3ForCausalLM" : _HfExamplesInfo ("google/gemma-3-1b-it" ,
148- min_transformers_version = "4.50" ),
148+ "Gemma3ForCausalLM" : _HfExamplesInfo ("google/gemma-3-1b-it" ),
149149 "GlmForCausalLM" : _HfExamplesInfo ("THUDM/glm-4-9b-chat-hf" ),
150150 "Glm4ForCausalLM" : _HfExamplesInfo (
151151 "THUDM/GLM-4-32B-0414" ,
152152 is_available_online = False ,
153153 min_transformers_version = "4.52.dev0"
154154 ),
155- "GPT2LMHeadModel" : _HfExamplesInfo ("gpt2" ),
156- "GPTBigCodeForCausalLM" : _HfExamplesInfo ("bigcode/starcoder" ),
157- "GPTJForCausalLM" : _HfExamplesInfo ("EleutherAI/gpt-j-6b" ),
158- "GPTNeoXForCausalLM" : _HfExamplesInfo ("EleutherAI/pythia-160m" ),
155+ "GPT2LMHeadModel" : _HfExamplesInfo ("openai-community/gpt2" ,
156+ {"alias" : "gpt2" }),
157+ "GPTBigCodeForCausalLM" : _HfExamplesInfo ("bigcode/starcoder" ,
158+ {"tiny" : "bigcode/tiny_starcoder_py" }), # noqa: E501
159+ "GPTJForCausalLM" : _HfExamplesInfo ("Milos/slovak-gpt-j-405M" ,
160+ {"6b" : "EleutherAI/gpt-j-6b" }),
161+ "GPTNeoXForCausalLM" : _HfExamplesInfo ("EleutherAI/pythia-70m" ,
162+ {"1b" : "EleutherAI/pythia-1.4b" }),
159163 "GraniteForCausalLM" : _HfExamplesInfo ("ibm/PowerLM-3b" ),
160164 "GraniteMoeForCausalLM" : _HfExamplesInfo ("ibm/PowerMoE-3b" ),
161- "GraniteMoeSharedForCausalLM" : _HfExamplesInfo ("ibm-research/moe-7b-1b-active-shared-experts" , # noqa: E501
162- min_transformers_version = "4.49" ), # noqa: E501
165+ "GraniteMoeSharedForCausalLM" : _HfExamplesInfo ("ibm-research/moe-7b-1b-active-shared-experts" ), # noqa: E501
163166 "Grok1ModelForCausalLM" : _HfExamplesInfo ("hpcai-tech/grok-1" ,
164167 trust_remote_code = True ),
165168 "InternLMForCausalLM" : _HfExamplesInfo ("internlm/internlm-chat-7b" ,
@@ -187,15 +190,17 @@ def check_available_online(
187190 "MiniMaxText01ForCausalLM" : _HfExamplesInfo ("MiniMaxAI/MiniMax-Text-01" ,
188191 trust_remote_code = True ),
189192 "MistralForCausalLM" : _HfExamplesInfo ("mistralai/Mistral-7B-Instruct-v0.1" ),
190- "MixtralForCausalLM" : _HfExamplesInfo ("mistralai/Mixtral-8x7B-Instruct-v0.1" ), # noqa: E501
193+ "MixtralForCausalLM" : _HfExamplesInfo ("mistralai/Mixtral-8x7B-Instruct-v0.1" , # noqa: E501
194+ {"falcon3" : "ehristoforu/Falcon3-MoE-2x7B-Insruct" }), # noqa: E501
191195 "QuantMixtralForCausalLM" : _HfExamplesInfo ("mistral-community/Mixtral-8x22B-v0.1-AWQ" ), # noqa: E501
192196 "MptForCausalLM" : _HfExamplesInfo ("mpt" , is_available_online = False ),
193197 "MPTForCausalLM" : _HfExamplesInfo ("mosaicml/mpt-7b" ),
194198 "NemotronForCausalLM" : _HfExamplesInfo ("nvidia/Minitron-8B-Base" ),
195199 "OlmoForCausalLM" : _HfExamplesInfo ("allenai/OLMo-1B-hf" ),
196200 "Olmo2ForCausalLM" : _HfExamplesInfo ("shanearora/OLMo-7B-1124-hf" ),
197201 "OlmoeForCausalLM" : _HfExamplesInfo ("allenai/OLMoE-1B-7B-0924-Instruct" ),
198- "OPTForCausalLM" : _HfExamplesInfo ("facebook/opt-iml-max-1.3b" ),
202+ "OPTForCausalLM" : _HfExamplesInfo ("facebook/opt-125m" ,
203+ {"1b" : "facebook/opt-iml-max-1.3b" }),
199204 "OrionForCausalLM" : _HfExamplesInfo ("OrionStarAI/Orion-14B-Chat" ,
200205 trust_remote_code = True ),
201206 "PersimmonForCausalLM" : _HfExamplesInfo ("adept/persimmon-8b-chat" ),
@@ -209,8 +214,8 @@ def check_available_online(
209214 trust_remote_code = True ),
210215 "QWenLMHeadModel" : _HfExamplesInfo ("Qwen/Qwen-7B-Chat" ,
211216 trust_remote_code = True ),
212- "Qwen2ForCausalLM" : _HfExamplesInfo ("Qwen/Qwen2-7B -Instruct" ,
213- extras = {"2.5" : "Qwen/Qwen2.5-7B -Instruct" }), # noqa: E501
217+ "Qwen2ForCausalLM" : _HfExamplesInfo ("Qwen/Qwen2-0.5B -Instruct" ,
218+ extras = {"2.5" : "Qwen/Qwen2.5-0.5B -Instruct" }), # noqa: E501
214219 "Qwen2MoeForCausalLM" : _HfExamplesInfo ("Qwen/Qwen1.5-MoE-A2.7B-Chat" ),
215220 "Qwen3ForCausalLM" : _HfExamplesInfo (
216221 "Qwen/Qwen3-8B" ,
@@ -236,8 +241,7 @@ def check_available_online(
236241 "XverseForCausalLM" : _HfExamplesInfo ("xverse/XVERSE-7B-Chat" ,
237242 is_available_online = False ,
238243 trust_remote_code = True ),
239- "Zamba2ForCausalLM" : _HfExamplesInfo ("Zyphra/Zamba2-7B-instruct" ,
240- min_transformers_version = "4.49" ),
244+ "Zamba2ForCausalLM" : _HfExamplesInfo ("Zyphra/Zamba2-7B-instruct" ),
241245 # [Encoder-decoder]
242246 "BartModel" : _HfExamplesInfo ("facebook/bart-base" ),
243247 "BartForConditionalGeneration" : _HfExamplesInfo ("facebook/bart-large-cnn" ),
@@ -280,9 +284,7 @@ def check_available_online(
280284 "BertForSequenceClassification" : _HfExamplesInfo ("cross-encoder/ms-marco-MiniLM-L-6-v2" ), # noqa: E501
281285 "RobertaForSequenceClassification" : _HfExamplesInfo ("cross-encoder/quora-roberta-base" ), # noqa: E501
282286 "XLMRobertaForSequenceClassification" : _HfExamplesInfo ("BAAI/bge-reranker-v2-m3" ), # noqa: E501
283- "ModernBertForSequenceClassification" :
284- _HfExamplesInfo ("Alibaba-NLP/gte-reranker-modernbert-base" ,
285- min_transformers_version = "4.49" ),
287+ "ModernBertForSequenceClassification" : _HfExamplesInfo ("Alibaba-NLP/gte-reranker-modernbert-base" ), # noqa: E501
286288}
287289
288290_MULTIMODAL_EXAMPLE_MODELS = {
@@ -298,8 +300,7 @@ def check_available_online(
298300 transformers_version_reason = "HF model is not compatible." , # noqa: E501
299301 hf_overrides = {"architectures" : ["DeepseekVLV2ForCausalLM" ]}), # noqa: E501
300302 "FuyuForCausalLM" : _HfExamplesInfo ("adept/fuyu-8b" ),
301- "Gemma3ForConditionalGeneration" : _HfExamplesInfo ("google/gemma-3-4b-it" ,
302- min_transformers_version = "4.50" ),
303+ "Gemma3ForConditionalGeneration" : _HfExamplesInfo ("google/gemma-3-4b-it" ),
303304 "GLM4VForCausalLM" : _HfExamplesInfo ("THUDM/glm-4v-9b" ,
304305 trust_remote_code = True ,
305306 hf_overrides = {"architectures" : ["GLM4VForCausalLM" ]}), # noqa: E501
@@ -335,7 +336,6 @@ def check_available_online(
335336 extras = {"2.6" : "openbmb/MiniCPM-V-2_6" }, # noqa: E501
336337 trust_remote_code = True ),
337338 "Mistral3ForConditionalGeneration" : _HfExamplesInfo ("mistralai/Mistral-Small-3.1-24B-Instruct-2503" , # noqa: E501
338- min_transformers_version = "4.50" , # noqa: E501
339339 extras = {"fp8" : "nm-testing/Mistral-Small-3.1-24B-Instruct-2503-FP8-dynamic" }), # noqa: E501
340340 "MolmoForCausalLM" : _HfExamplesInfo ("allenai/Molmo-7B-D-0924" ,
341341 max_transformers_version = "4.48" ,
@@ -361,8 +361,7 @@ def check_available_online(
361361 hf_overrides = {"architectures" : ["QwenVLForConditionalGeneration" ]}), # noqa: E501
362362 "Qwen2AudioForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2-Audio-7B-Instruct" ), # noqa: E501
363363 "Qwen2VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2-VL-2B-Instruct" ), # noqa: E501
364- "Qwen2_5_VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-VL-3B-Instruct" , # noqa: E501
365- min_transformers_version = "4.49" ), # noqa: E501
364+ "Qwen2_5_VLForConditionalGeneration" : _HfExamplesInfo ("Qwen/Qwen2.5-VL-3B-Instruct" ), # noqa: E501
366365 "Qwen2_5OmniModel" : _HfExamplesInfo ("Qwen/Qwen2.5-Omni-7B" , # noqa: E501
367366 min_transformers_version = "4.52" ), # noqa: E501
368367 "SkyworkR1VChatModel" : _HfExamplesInfo ("Skywork/Skywork-R1V-38B" ),
0 commit comments