You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
<p>KTransformers is a research project focused on efficient inference and fine-tuning of large language models through CPU-GPU heterogeneous computing. The project has evolved into <strong>two core modules</strong>: <ahref="https://github.com/kvcache-ai/ktransformers/tree/main/kt-kernel/">kt-kernel</a> and <ahref="https://github.com/kvcache-ai/ktransformers/tree/main/kt-sft">kt-sft</a>.</p>
<li><strong>Dec 24, 2025</strong>: Support Native MiniMax-M2.1 inference. (<ahref="./doc/en/MiniMax-M2.1-Tutorial.html">Tutorial</a>)</li>
201
+
<li><strong>Dec 24, 2025</strong>: Support Native MiniMax-M2.1 inference. (<ahref="./doc/en/kt-kernel/MiniMax-M2.1-Tutorial.html">Tutorial</a>)</li>
202
202
<li><strong>Dec 22, 2025</strong>: Support RL-DPO fine-tuning with LLaMA-Factory. (<ahref="./doc/en/SFT/DPO_tutorial.html">Tutorial</a>)</li>
203
-
<li><strong>Dec 5, 2025</strong>: Support Native Kimi-K2-Thinking inference (<ahref="./doc/en/Kimi-K2-Thinking-Native.html">Tutorial</a>)</li>
203
+
<li><strong>Dec 5, 2025</strong>: Support Native Kimi-K2-Thinking inference (<ahref="./doc/en/kt-kernel/Kimi-K2-Thinking-Native.html">Tutorial</a>)</li>
204
204
<li><strong>Nov 6, 2025</strong>: Support Kimi-K2-Thinking inference (<ahref="./doc/en/Kimi-K2-Thinking.html">Tutorial</a>) and fine-tune (<ahref="./doc/en/SFT_Installation_Guide_KimiK2.html">Tutorial</a>)</li>
<p>KTransformers is a research project focused on efficient inference and fine-tuning of large language models through CPU-GPU heterogeneous computing. The project has evolved into <strong>two core modules</strong>: <ahref="https://github.com/kvcache-ai/ktransformers/tree/main/kt-kernel/">kt-kernel</a> and <ahref="https://github.com/kvcache-ai/ktransformers/tree/main/kt-sft">kt-sft</a>.</p>
<li><strong>Dec 24, 2025</strong>: Support Native MiniMax-M2.1 inference. (<ahref="doc/en/MiniMax-M2.1-Tutorial.html">Tutorial</a>)</li>
200
+
<li><strong>Dec 24, 2025</strong>: Support Native MiniMax-M2.1 inference. (<ahref="doc/en/kt-kernel/MiniMax-M2.1-Tutorial.html">Tutorial</a>)</li>
201
201
<li><strong>Dec 22, 2025</strong>: Support RL-DPO fine-tuning with LLaMA-Factory. (<ahref="doc/en/SFT/DPO_tutorial.html">Tutorial</a>)</li>
202
-
<li><strong>Dec 5, 2025</strong>: Support Native Kimi-K2-Thinking inference (<ahref="doc/en/Kimi-K2-Thinking-Native.html">Tutorial</a>)</li>
202
+
<li><strong>Dec 5, 2025</strong>: Support Native Kimi-K2-Thinking inference (<ahref="doc/en/kt-kernel/Kimi-K2-Thinking-Native.html">Tutorial</a>)</li>
203
203
<li><strong>Nov 6, 2025</strong>: Support Kimi-K2-Thinking inference (<ahref="doc/en/Kimi-K2-Thinking.html">Tutorial</a>) and fine-tune (<ahref="doc/en/SFT_Installation_Guide_KimiK2.html">Tutorial</a>)</li>
0 commit comments