@@ -76,12 +76,12 @@ def cross_entropy_forward(self, inputs: torch.Tensor, target: torch.Tensor) -> t
76
76
ModelMeta (
77
77
LLMModelType .chatglm2 , [
78
78
ModelGroup ([
79
- Model ('ZhipuAI/chatglm2-6b' , 'THUDM /chatglm2-6b' ),
80
- Model ('ZhipuAI/chatglm2-6b-32k' , 'THUDM /chatglm2-6b-32k' )
79
+ Model ('ZhipuAI/chatglm2-6b' , 'zai-org /chatglm2-6b' ),
80
+ Model ('ZhipuAI/chatglm2-6b-32k' , 'zai-org /chatglm2-6b-32k' )
81
81
],
82
82
requires = ['transformers<4.42' ]),
83
83
ModelGroup (
84
- [Model ('ZhipuAI/codegeex2-6b' , 'THUDM /codegeex2-6b' )],
84
+ [Model ('ZhipuAI/codegeex2-6b' , 'zai-org /codegeex2-6b' )],
85
85
requires = ['transformers<4.34' ],
86
86
tags = ['coding' ],
87
87
),
@@ -95,10 +95,10 @@ def cross_entropy_forward(self, inputs: torch.Tensor, target: torch.Tensor) -> t
95
95
ModelMeta (
96
96
LLMModelType .chatglm3 , [
97
97
ModelGroup ([
98
- Model ('ZhipuAI/chatglm3-6b' , 'THUDM /chatglm3-6b' ),
99
- Model ('ZhipuAI/chatglm3-6b-base' , 'THUDM /chatglm3-6b-base' ),
100
- Model ('ZhipuAI/chatglm3-6b-32k' , 'THUDM /chatglm3-6b-32k' ),
101
- Model ('ZhipuAI/chatglm3-6b-128k' , 'THUDM /chatglm3-6b-128k' ),
98
+ Model ('ZhipuAI/chatglm3-6b' , 'zai-org /chatglm3-6b' ),
99
+ Model ('ZhipuAI/chatglm3-6b-base' , 'zai-org /chatglm3-6b-base' ),
100
+ Model ('ZhipuAI/chatglm3-6b-32k' , 'zai-org /chatglm3-6b-32k' ),
101
+ Model ('ZhipuAI/chatglm3-6b-128k' , 'zai-org /chatglm3-6b-128k' ),
102
102
])
103
103
],
104
104
TemplateType .glm4 ,
@@ -128,12 +128,12 @@ def get_model_tokenizer_glm4(model_dir: str,
128
128
LLMModelType .glm4 ,
129
129
[
130
130
ModelGroup ([
131
- Model ('ZhipuAI/glm-4-9b-chat' , 'THUDM /glm-4-9b-chat' ),
132
- Model ('ZhipuAI/glm-4-9b' , 'THUDM /glm-4-9b' ),
133
- Model ('ZhipuAI/glm-4-9b-chat-1m' , 'THUDM /glm-4-9b-chat-1m' ),
131
+ Model ('ZhipuAI/glm-4-9b-chat' , 'zai-org /glm-4-9b-chat' ),
132
+ Model ('ZhipuAI/glm-4-9b' , 'zai-org /glm-4-9b' ),
133
+ Model ('ZhipuAI/glm-4-9b-chat-1m' , 'zai-org /glm-4-9b-chat-1m' ),
134
134
]),
135
135
ModelGroup ([
136
- Model ('ZhipuAI/LongWriter-glm4-9b' , 'THUDM /LongWriter-glm4-9b' ),
136
+ Model ('ZhipuAI/LongWriter-glm4-9b' , 'zai-org /LongWriter-glm4-9b' ),
137
137
])
138
138
],
139
139
TemplateType .glm4 ,
@@ -148,11 +148,11 @@ def get_model_tokenizer_glm4(model_dir: str,
148
148
LLMModelType .glm4_0414 ,
149
149
[
150
150
ModelGroup ([
151
- Model ('ZhipuAI/GLM-4-9B-0414' , 'THUDM /GLM-4-9B-0414' ),
152
- Model ('ZhipuAI/GLM-4-32B-0414' , 'THUDM /GLM-4-32B-0414' ),
153
- Model ('ZhipuAI/GLM-4-32B-Base-0414' , 'THUDM /GLM-4-32B-Base-0414' ),
154
- Model ('ZhipuAI/GLM-Z1-9B-0414' , 'THUDM /GLM-Z1-9B-0414' ),
155
- Model ('ZhipuAI/GLM-Z1-32B-0414' , 'THUDM /GLM-Z1-32B-0414' ),
151
+ Model ('ZhipuAI/GLM-4-9B-0414' , 'zai-org /GLM-4-9B-0414' ),
152
+ Model ('ZhipuAI/GLM-4-32B-0414' , 'zai-org /GLM-4-32B-0414' ),
153
+ Model ('ZhipuAI/GLM-4-32B-Base-0414' , 'zai-org /GLM-4-32B-Base-0414' ),
154
+ Model ('ZhipuAI/GLM-Z1-9B-0414' , 'zai-org /GLM-Z1-9B-0414' ),
155
+ Model ('ZhipuAI/GLM-Z1-32B-0414' , 'zai-org /GLM-Z1-32B-0414' ),
156
156
])
157
157
],
158
158
TemplateType .glm4_0414 ,
@@ -166,7 +166,7 @@ def get_model_tokenizer_glm4(model_dir: str,
166
166
ModelMeta (
167
167
LLMModelType .glm4_z1_rumination ,
168
168
[ModelGroup ([
169
- Model ('ZhipuAI/GLM-Z1-Rumination-32B-0414' , 'THUDM /GLM-Z1-Rumination-32B-0414' ),
169
+ Model ('ZhipuAI/GLM-Z1-Rumination-32B-0414' , 'zai-org /GLM-Z1-Rumination-32B-0414' ),
170
170
])],
171
171
TemplateType .glm4_z1_rumination ,
172
172
get_model_tokenizer_with_flash_attn ,
@@ -179,7 +179,7 @@ def get_model_tokenizer_glm4(model_dir: str,
179
179
ModelMeta (
180
180
LLMModelType .longwriter_llama3_1 ,
181
181
[ModelGroup ([
182
- Model ('ZhipuAI/LongWriter-llama3.1-8b' , 'THUDM /LongWriter-llama3.1-8b' ),
182
+ Model ('ZhipuAI/LongWriter-llama3.1-8b' , 'zai-org /LongWriter-llama3.1-8b' ),
183
183
])],
184
184
TemplateType .longwriter_llama ,
185
185
get_model_tokenizer_with_flash_attn ,
@@ -192,7 +192,7 @@ def get_model_tokenizer_glm4(model_dir: str,
192
192
ModelMeta (
193
193
LLMModelType .codegeex4 ,
194
194
[ModelGroup ([
195
- Model ('ZhipuAI/codegeex4-all-9b' , 'THUDM /codegeex4-all-9b' ),
195
+ Model ('ZhipuAI/codegeex4-all-9b' , 'zai-org /codegeex4-all-9b' ),
196
196
])],
197
197
TemplateType .codegeex4 ,
198
198
get_model_tokenizer_glm4 ,
@@ -231,13 +231,13 @@ def get_model_tokenizer_glm4v(model_dir: str,
231
231
[
232
232
ModelGroup (
233
233
[
234
- Model ('ZhipuAI/glm-4v-9b' , 'THUDM /glm-4v-9b' ),
234
+ Model ('ZhipuAI/glm-4v-9b' , 'zai-org /glm-4v-9b' ),
235
235
],
236
236
requires = ['transformers>=4.42,<4.45' ],
237
237
),
238
238
ModelGroup (
239
239
[
240
- Model ('ZhipuAI/cogagent-9b-20241220' , 'THUDM /cogagent-9b-20241220' ),
240
+ Model ('ZhipuAI/cogagent-9b-20241220' , 'zai-org /cogagent-9b-20241220' ),
241
241
],
242
242
requires = ['transformers>=4.42' ],
243
243
)
@@ -268,8 +268,8 @@ def get_model_tokenizer_glm4_1v(*args, **kwargs):
268
268
[
269
269
ModelGroup (
270
270
[
271
- Model ('ZhipuAI/GLM-4.1V-9B-Base' , 'THUDM /GLM-4.1V-9B-Base' ),
272
- Model ('ZhipuAI/GLM-4.1V-9B-Thinking' , 'THUDM /GLM-4.1V-9B-Thinking' ),
271
+ Model ('ZhipuAI/GLM-4.1V-9B-Base' , 'zai-org /GLM-4.1V-9B-Base' ),
272
+ Model ('ZhipuAI/GLM-4.1V-9B-Thinking' , 'zai-org /GLM-4.1V-9B-Thinking' ),
273
273
],
274
274
requires = ['transformers>=4.53' ],
275
275
),
@@ -301,7 +301,7 @@ def get_model_tokenizer_cogvlm(model_dir: str,
301
301
ModelMeta (
302
302
MLLMModelType .cogvlm , [
303
303
ModelGroup ([
304
- Model ('ZhipuAI/cogvlm-chat' , 'THUDM /cogvlm-chat-hf' ),
304
+ Model ('ZhipuAI/cogvlm-chat' , 'zai-org /cogvlm-chat-hf' ),
305
305
]),
306
306
],
307
307
TemplateType .cogvlm ,
@@ -314,7 +314,7 @@ def get_model_tokenizer_cogvlm(model_dir: str,
314
314
ModelMeta (
315
315
MLLMModelType .cogagent_chat , [
316
316
ModelGroup ([
317
- Model ('ZhipuAI/cogagent-chat' , 'THUDM /cogagent-chat-hf' ),
317
+ Model ('ZhipuAI/cogagent-chat' , 'zai-org /cogagent-chat-hf' ),
318
318
]),
319
319
],
320
320
TemplateType .cogagent_chat ,
@@ -326,7 +326,7 @@ def get_model_tokenizer_cogvlm(model_dir: str,
326
326
register_model (
327
327
ModelMeta (
328
328
MLLMModelType .cogagent_vqa , [ModelGroup ([
329
- Model ('ZhipuAI/cogagent-vqa' , 'THUDM /cogagent-vqa-hf' ),
329
+ Model ('ZhipuAI/cogagent-vqa' , 'zai-org /cogagent-vqa-hf' ),
330
330
])],
331
331
TemplateType .cogagent_vqa ,
332
332
get_model_tokenizer_cogvlm ,
@@ -353,8 +353,8 @@ def get_model_tokenizer_cogvlm2(*args, **kwargs):
353
353
ModelMeta (
354
354
MLLMModelType .cogvlm2 , [
355
355
ModelGroup ([
356
- Model ('ZhipuAI/cogvlm2-llama3-chat-19B' , 'THUDM /cogvlm2-llama3-chat-19B' ),
357
- Model ('ZhipuAI/cogvlm2-llama3-chinese-chat-19B' , 'THUDM /cogvlm2-llama3-chinese-chat-19B' ),
356
+ Model ('ZhipuAI/cogvlm2-llama3-chat-19B' , 'zai-org /cogvlm2-llama3-chat-19B' ),
357
+ Model ('ZhipuAI/cogvlm2-llama3-chinese-chat-19B' , 'zai-org /cogvlm2-llama3-chinese-chat-19B' ),
358
358
]),
359
359
],
360
360
TemplateType .cogvlm2 ,
@@ -368,7 +368,7 @@ def get_model_tokenizer_cogvlm2(*args, **kwargs):
368
368
MLLMModelType .cogvlm2_video ,
369
369
[
370
370
ModelGroup ([
371
- Model ('ZhipuAI/cogvlm2-video-llama3-chat' , 'THUDM /cogvlm2-video-llama3-chat' ),
371
+ Model ('ZhipuAI/cogvlm2-video-llama3-chat' , 'zai-org /cogvlm2-video-llama3-chat' ),
372
372
]),
373
373
],
374
374
TemplateType .cogvlm2_video ,
@@ -384,8 +384,8 @@ def get_model_tokenizer_cogvlm2(*args, **kwargs):
384
384
LLMModelType .glm_edge ,
385
385
[
386
386
ModelGroup ([
387
- Model ('ZhipuAI/glm-edge-1.5b-chat' , 'THUDM /glm-edge-1.5b-chat' ),
388
- Model ('ZhipuAI/glm-edge-4b-chat' , 'THUDM /glm-edge-4b-chat' ),
387
+ Model ('ZhipuAI/glm-edge-1.5b-chat' , 'zai-org /glm-edge-1.5b-chat' ),
388
+ Model ('ZhipuAI/glm-edge-4b-chat' , 'zai-org /glm-edge-4b-chat' ),
389
389
]),
390
390
],
391
391
TemplateType .glm4 ,
@@ -408,8 +408,8 @@ def get_model_tokenizer_glm_edge_v(model_dir: str, *args, **kwargs):
408
408
MLLMModelType .glm_edge_v ,
409
409
[
410
410
ModelGroup ([
411
- Model ('ZhipuAI/glm-edge-v-2b' , 'THUDM /glm-edge-v-2b' ),
412
- Model ('ZhipuAI/glm-edge-4b-chat' , 'THUDM /glm-edge-4b-chat' ),
411
+ Model ('ZhipuAI/glm-edge-v-2b' , 'zai-org /glm-edge-v-2b' ),
412
+ Model ('ZhipuAI/glm-edge-4b-chat' , 'zai-org /glm-edge-4b-chat' ),
413
413
]),
414
414
],
415
415
TemplateType .glm_edge_v ,
@@ -425,16 +425,39 @@ def get_model_tokenizer_glm_edge_v(model_dir: str, *args, **kwargs):
425
425
LLMModelType .glm4_5 ,
426
426
[
427
427
ModelGroup ([
428
- Model ('ZhipuAI/GLM-4.5-Air-Base' , 'THUDM /GLM-4.5-Air-Base' ),
429
- Model ('ZhipuAI/GLM-4.5-Air' , 'THUDM /GLM-4.5-Air' ),
430
- Model ('ZhipuAI/GLM-4.5-Air-FP8' , 'THUDM /GLM-4.5-Air-FP8' ),
431
- Model ('ZhipuAI/GLM-4.5-Base' , 'THUDM /GLM-4.5-Base' ),
432
- Model ('ZhipuAI/GLM-4.5' , 'THUDM /GLM-4.5' ),
433
- Model ('ZhipuAI/GLM-4.5-FP8' , 'THUDM /GLM-4.5-FP8' ),
428
+ Model ('ZhipuAI/GLM-4.5-Air-Base' , 'zai-org /GLM-4.5-Air-Base' ),
429
+ Model ('ZhipuAI/GLM-4.5-Air' , 'zai-org /GLM-4.5-Air' ),
430
+ Model ('ZhipuAI/GLM-4.5-Air-FP8' , 'zai-org /GLM-4.5-Air-FP8' ),
431
+ Model ('ZhipuAI/GLM-4.5-Base' , 'zai-org /GLM-4.5-Base' ),
432
+ Model ('ZhipuAI/GLM-4.5' , 'zai-org /GLM-4.5' ),
433
+ Model ('ZhipuAI/GLM-4.5-FP8' , 'zai-org /GLM-4.5-FP8' ),
434
434
]),
435
435
],
436
436
TemplateType .glm4_5 ,
437
437
get_model_tokenizer_with_flash_attn ,
438
438
architectures = ['Glm4MoeForCausalLM' ],
439
439
requires = ['transformers>=4.54' ],
440
440
))
441
+
442
+
443
+ def get_model_tokenizer_glm4_5v (* args , ** kwargs ):
444
+ from transformers import Glm4vMoeForConditionalGeneration
445
+ kwargs ['automodel_class' ] = kwargs ['automodel_class' ] or Glm4vMoeForConditionalGeneration
446
+ return get_model_tokenizer_multimodal (* args , ** kwargs )
447
+
448
+
449
+ register_model (
450
+ ModelMeta (
451
+ MLLMModelType .glm4_5v ,
452
+ [
453
+ ModelGroup ([
454
+ Model ('ZhipuAI/GLM-4.5V' , 'zai-org/GLM-4.5V' ),
455
+ Model ('ZhipuAI/GLM-4.5V-FP8' , 'zai-org/GLM-4.5V-FP8' ),
456
+ ]),
457
+ ],
458
+ TemplateType .glm4_5v ,
459
+ get_model_tokenizer_glm4_5v ,
460
+ model_arch = ModelArch .glm4_1v ,
461
+ architectures = ['Glm4vMoeForConditionalGeneration' ],
462
+ requires = ['transformers>=4.56.0.dev' ],
463
+ ))
0 commit comments