You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
"It's optimized for an assistant-like chat use cases, with native support for function calling.\n"+
151
151
"This version of the model utilizes a Mixture of Experts architecture, with only 3B active parameters, thus making it very fast.\n"+
152
152
"Mixtures of Experts (MoE) is a technique where different models, each skilled in solving a particular kind of problem, work together to the improve the overall performance on complex tasks.\n"+
153
-
"This model is censored, but its responses quality on many topics is extremely high.\n"+
153
+
"This model is censored, but its responses quality on many topics is high compared to its high generation speed.\n"+
154
154
"This is the 30 billion parameters Mixtures of Experts (MoE) version of the model.\n"+
155
155
"Its performance is comparable and even surpasses DeepSeek V3 and GPT-4o.",
0 commit comments