Skip to content

Create Specialized LLM Models for the Lightweight LLM Server for Integration TestΒ #239

@rootfs

Description

@rootfs

Is your feature request related to a problem? Please describe.
Since the router is to support Mixture-of-Models, we need to have integration test suites to verify such functionalities.

Now with #228, we are able to use a real yet lightweight models as LLM server, we can extend such server to support multiple and specialized models.

Describe the solution you'd like

  • Fine tune the lightweight models, Qwen3 0.6B or Gemma 270m, for different specializations.
  • Run the entire model_eval and config generation workflow
  • Run benchmark to evaluate the router's performance and accuracy

@Xunzhuo @yossiovadia @yuluo-yx @JaredforReal WDYT?

Metadata

Metadata

Assignees

No one assigned

    Projects

    No projects

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions