Skip to content

Commit 2ba7502

Browse files
Add GLM Export (#51)
Add ONNX export for GLM
1 parent 5ab5bf3 commit 2ba7502

File tree

7 files changed

+15
-2
lines changed

7 files changed

+15
-2
lines changed

docs/source/onnx/overview.mdx

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -52,6 +52,7 @@ Supported architectures from [🤗 Transformers](https://huggingface.co/docs/tra
5252
- ESM
5353
- Falcon
5454
- Flaubert
55+
- GLM
5556
- GPT-2
5657
- GPT-BigCode
5758
- GPT-J

optimum/exporters/onnx/model_configs.py

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -456,6 +456,11 @@ class CohereOnnxConfig(LlamaOnnxConfig):
456456
_MODEL_PATCHER = CohereModelPatcher
457457

458458

459+
@register_tasks_manager_onnx("glm", *COMMON_TEXT_GENERATION_TASKS)
460+
class GLMOnnxConfig(LlamaOnnxConfig):
461+
MIN_TRANSFORMERS_VERSION = version.parse("4.46.0")
462+
463+
459464
@register_tasks_manager_onnx("helium", *COMMON_TEXT_GENERATION_TASKS)
460465
class HeliumOnnxConfig(LlamaOnnxConfig):
461466
MIN_TRANSFORMERS_VERSION = version.parse("4.49.0")

optimum/exporters/onnx/utils.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -75,12 +75,14 @@
7575
"cohere",
7676
"falcon",
7777
"gemma",
78+
"glm",
7879
"gpt2",
7980
"gpt_bigcode",
8081
"gpt_neo",
8182
"gpt_neox",
8283
"gpt_oss",
8384
"gptj",
85+
"granite",
8486
"helium",
8587
"imagegpt",
8688
"internlm2",
@@ -92,7 +94,6 @@
9294
"qwen2",
9395
"qwen3",
9496
"qwen3_moe",
95-
"granite",
9697
"smollm3",
9798
"stablelm",
9899
"olmo2",

optimum/onnxruntime/modeling_decoder.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -246,6 +246,8 @@ def __init__(
246246
"deepseek_v3",
247247
"cohere",
248248
"gemma",
249+
"glm",
250+
"granite",
249251
"gpt_oss",
250252
"helium",
251253
"mistral",
@@ -254,7 +256,6 @@ def __init__(
254256
"qwen2",
255257
"qwen3",
256258
"qwen3_moe",
257-
"granite",
258259
"smollm3",
259260
"stablelm",
260261
}:

tests/exporters/onnx/utils_tests.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -103,6 +103,7 @@
103103
},
104104
"flaubert": "hf-internal-testing/tiny-random-flaubert",
105105
"gemma": "fxmarty/tiny-random-GemmaForCausalLM",
106+
"glm": "hf-internal-testing/tiny-random-GlmForCausalLM",
106107
"glpn": "hf-internal-testing/tiny-random-GLPNModel",
107108
"gpt2": "hf-internal-testing/tiny-random-gpt2",
108109
"gpt_bigcode": "hf-internal-testing/tiny-random-GPTBigCodeModel",

tests/onnxruntime/test_decoder.py

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -33,6 +33,7 @@
3333
CohereOnnxConfig,
3434
DeepSeekV3OnnxConfig,
3535
GemmaOnnxConfig,
36+
GLMOnnxConfig,
3637
GPTOssOnnxConfig,
3738
GraniteOnnxConfig,
3839
HeliumOnnxConfig,
@@ -112,6 +113,8 @@ class ORTModelForCausalLMIntegrationTest(ORTModelTestMixin):
112113
SUPPORTED_ARCHITECTURES.append("qwen2")
113114
if is_transformers_version(">=", str(GemmaOnnxConfig.MIN_TRANSFORMERS_VERSION)):
114115
SUPPORTED_ARCHITECTURES.append("gemma")
116+
if is_transformers_version(">=", str(GLMOnnxConfig.MIN_TRANSFORMERS_VERSION)):
117+
SUPPORTED_ARCHITECTURES.append("glm")
115118
if is_transformers_version(">=", str(MPTOnnxConfig.MIN_TRANSFORMERS_VERSION)):
116119
SUPPORTED_ARCHITECTURES.append("mpt")
117120
if is_transformers_version(">=", str(NemotronOnnxConfig.MIN_TRANSFORMERS_VERSION)):

tests/onnxruntime/testing_utils.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -68,6 +68,7 @@
6868
"flaubert": "hf-internal-testing/tiny-random-flaubert",
6969
"flux": "optimum-internal-testing/tiny-random-flux",
7070
"gemma": "fxmarty/tiny-random-GemmaForCausalLM",
71+
"glm": "hf-internal-testing/tiny-random-GlmForCausalLM",
7172
"gpt2": "hf-internal-testing/tiny-random-GPT2LMHeadModel",
7273
"gpt_bigcode": "hf-internal-testing/tiny-random-GPTBigCodeModel",
7374
"gpt_bigcode-multi_query-False": "optimum-internal-testing/tiny-random-gpt_bigcode-multi_query-False",

0 commit comments

Comments
 (0)