You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: _posts/2025-09-01-semantic-router.md
+4-6Lines changed: 4 additions & 6 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -36,8 +36,6 @@ Other companies are adopting similar strategies:
36
36
37
37
In short: the industry is entering an era where **no token should be wasted**.
38
38
39
-
---
40
-
41
39
## Recent Research: vLLM Semantic Router
42
40
43
41
Amid this shift toward hybrid reasoning, we focus on the **open-source inference engine vLLM**.
@@ -72,9 +70,9 @@ In knowledge-intensive areas such as business and economics, accuracy improvemen
72
70
73
71
The Semantic Router is not the isolated result of a single paper but a collaborative outcome of sustained community contributions:
74
72
75
-
* Originally proposed by **Dr. Chen Huamin**, Distinguished Engineer at **Red Hat**, in early **2025** across multiple open-source communities.
76
-
* Iterated and further developed by **Xunzhuo Liu** at **Tencent**, later contributed to the vLLM community.
77
-
***Dr. Wang Chen** from **IBM Research** and **Dr. Chen Huamin** will present the project at **KubeCon North America 2025**.
73
+
* Originally proposed by **[Dr. Chen Huamin](https://www.linkedin.com/in/huaminchen)**, Distinguished Engineer at **Red Hat**, in early **2025** across multiple open-source communities.
74
+
* Iterated and further developed by **[Xunzhuo Liu](https://www.linkedin.com/in/bitliu)** at **Tencent**, later contributed to the vLLM community.
75
+
***[Dr. Wang Chen](https://www.linkedin.com/in/chenw615)** from **IBM Research** and **Dr. Chen Huamin** will present the project at **KubeCon North America 2025**.
78
76
79
77
The mission is clear: to serve as an **inference accelerator** for open-source large models:
80
78
@@ -84,7 +82,7 @@ The mission is clear: to serve as an **inference accelerator** for open-source l
84
82
85
83
The vLLM Semantic Router is therefore not just a research milestone but an **essential bridge for open-source AI infrastructure**, translating **academic innovation into industrial application**.
86
84
87
-
You can start exploring the project here: [https://github.com/vllm-project/semantic-router](https://github.com/vllm-project/semantic-router).
85
+
You can start exploring the project at [https://github.com/vllm-project/semantic-router](https://github.com/vllm-project/semantic-router). We're currently working on the [v0.1 Roadmap](https://github.com/vllm-project/semantic-router/issues/14) and have established a [Work Group](https://github.com/vllm-project/semantic-router/issues/15). We welcome your thoughts and invite you to join us!
0 commit comments