### Feature Description Minimax M2 is a new competitor to GLM 4.6, with a strong focus on agent-related capabilities, and It seems a MOE-based model. ### Motivation Maybe the MoE optimization implemented in ik_llama.cpp could provide better performance than the upstream llama.cpp? ### Possible Implementation There’s already a pending upstream PR in llama.cpp: [#16831](https://github.com/ggml-org/llama.cpp/pull/16831) Would it be feasible to add support by merging that PR and adjusting any conflicts?