You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: _posts/2025-08-15-glm45-vllm.md
+17-21Lines changed: 17 additions & 21 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,15 +1,15 @@
1
1
---
2
2
layout: post
3
-
title: "Use vLLM to deploy GLM-4.5 and GLM-4.5V model"
3
+
title: "GLM-4.5 Meets vLLM: Built for Intelligent Agents"
4
4
author: "Yuxuan Zhang"
5
5
image: /assets/logos/vllm-logo-text-light.png
6
6
---
7
7
8
-
#Use vLLM to deploy GLM-4.5 and GLM-4.5V model
8
+
## Introduction
9
9
10
-
## Model Introduction
10
+
[General Language Model (GLM)](https://aclanthology.org/2022.acl-long.26/) is a family of foundation models created by Zhipu.ai (now renamed to [Z.ai](https://z.ai/)). The GLM team has long-term collaboration with vLLM team, dating back to the early days of vLLM and the popular [ChatGLM model series](https://github.com/zai-org/ChatGLM-6B). Recently, the GLM team released the GLM-4.5 and GLM-4.5V model series, which are designed for intelligent agents. They are the top trending models in Huggingface model hub right now.
11
11
12
-
The GLM-4.5 series models are foundation models designed for intelligent agents. GLM-4.5 has 355 billion total
12
+
GLM-4.5 has 355 billion total
13
13
parameters with 32 billion active parameters, while GLM-4.5-Air adopts a more compact design with 106 billion total
14
14
parameters and 12 billion active parameters. GLM-4.5 models unify reasoning, coding, and intelligent agent capabilities
15
15
to meet the complex demands of intelligent agent applications.
@@ -28,10 +28,10 @@ among models of the same scale on 42 public vision-language benchmarks.
0 commit comments