Skip to content
This repository was archived by the owner on Jan 16, 2026. It is now read-only.

Commit 6ba46db

Browse files
committed
chore: limit recommended models to working implementations
Only Qwen2 has proper RoPE support currently. Other models (Phi3, Gemma3n, Mistral) need their implementations fixed to apply RoPE. - Qwen2.5: Working with non-quantized models - Llama: Working but requires HuggingFace authentication - Others: Commented out until RoPE implementations are fixed
1 parent d6c9d5f commit 6ba46db

File tree

1 file changed

+13
-29
lines changed

1 file changed

+13
-29
lines changed

packages/node-mlx/src/index.ts

Lines changed: 13 additions & 29 deletions
Original file line numberDiff line numberDiff line change
@@ -153,41 +153,25 @@ export interface Model {
153153
// MARK: - Recommended Models
154154

155155
export const RECOMMENDED_MODELS = {
156-
// Phi (Microsoft) - Efficient reasoning
157-
phi4: "mlx-community/Phi-4-mini-instruct-4bit",
158-
"phi-4": "mlx-community/Phi-4-mini-instruct-4bit",
159-
phi3: "mlx-community/Phi-3-mini-4k-instruct-4bit",
160-
"phi-3": "mlx-community/Phi-3-mini-4k-instruct-4bit",
161-
"phi-3-mini": "mlx-community/Phi-3-mini-4k-instruct-4bit",
162-
phi: "mlx-community/Phi-4-mini-instruct-4bit", // Default to latest
163-
164-
// Gemma 3n (Google) - Efficient on-device model
165-
gemma3n: "mlx-community/gemma-3n-E4B-it-lm-4bit",
166-
"gemma-3n": "mlx-community/gemma-3n-E4B-it-lm-4bit",
167-
"gemma-3n-2b": "mlx-community/gemma-3n-E2B-it-lm-4bit",
168-
"gemma-3n-4b": "mlx-community/gemma-3n-E4B-it-lm-4bit",
169-
gemma: "mlx-community/gemma-3n-E4B-it-lm-4bit", // Default to latest
170-
171-
// Llama 3.2 (Meta) - Fast and capable
172-
// Note: Using non-quantized models until quantized weight loading is fixed
173-
llama: "meta-llama/Llama-3.2-1B-Instruct",
174-
"llama-3.2": "meta-llama/Llama-3.2-1B-Instruct",
175-
"llama-3.2-1b": "meta-llama/Llama-3.2-1B-Instruct",
176-
"llama-3.2-3b": "meta-llama/Llama-3.2-3B-Instruct",
177-
178-
// Qwen 2.5/3 (Alibaba) - Great multilingual support
179-
// Note: Using non-quantized models until quantized weight loading is fixed
156+
// Qwen 2.5 (Alibaba) - RECOMMENDED: Working with proper RoPE support
157+
// Using non-quantized models - quantized models have loading issues
180158
qwen: "Qwen/Qwen2.5-1.5B-Instruct",
181159
"qwen-2.5": "Qwen/Qwen2.5-1.5B-Instruct",
182160
"qwen-2.5-0.5b": "Qwen/Qwen2.5-0.5B-Instruct",
183161
"qwen-2.5-1.5b": "Qwen/Qwen2.5-1.5B-Instruct",
184162
"qwen-2.5-3b": "Qwen/Qwen2.5-3B-Instruct",
185-
qwen3: "Qwen/Qwen3-4B",
186-
"qwen-3": "Qwen/Qwen3-4B",
187163

188-
// Mistral/Ministral
189-
mistral: "mlx-community/Mistral-7B-Instruct-v0.3-4bit",
190-
ministral: "mlx-community/Ministral-8B-Instruct-2410-4bit"
164+
// Llama 3.2 (Meta) - Requires HuggingFace authentication
165+
// Note: meta-llama models require accepting license at huggingface.co
166+
llama: "meta-llama/Llama-3.2-1B-Instruct",
167+
"llama-3.2": "meta-llama/Llama-3.2-1B-Instruct",
168+
"llama-3.2-1b": "meta-llama/Llama-3.2-1B-Instruct",
169+
"llama-3.2-3b": "meta-llama/Llama-3.2-3B-Instruct"
170+
171+
// TODO: These models need RoPE fixes in their implementations:
172+
// - Phi3/Phi4: Missing RoPE application
173+
// - Gemma3n: MoE config parsing + RoPE
174+
// - Mistral/Ministral: Missing RoPE application
191175
} as const
192176

193177
export type RecommendedModelKey = keyof typeof RECOMMENDED_MODELS

0 commit comments

Comments
 (0)