@@ -153,25 +153,30 @@ export interface Model {
153153// MARK: - Recommended Models
154154
155155export const RECOMMENDED_MODELS = {
156- // Qwen 2.5 (Alibaba) - RECOMMENDED: Working with proper RoPE support
156+ // Qwen 2.5 (Alibaba) - Working with proper RoPE support
157157 // Using non-quantized models - quantized models have loading issues
158158 qwen : "Qwen/Qwen2.5-1.5B-Instruct" ,
159159 "qwen-2.5" : "Qwen/Qwen2.5-1.5B-Instruct" ,
160160 "qwen-2.5-0.5b" : "Qwen/Qwen2.5-0.5B-Instruct" ,
161161 "qwen-2.5-1.5b" : "Qwen/Qwen2.5-1.5B-Instruct" ,
162162 "qwen-2.5-3b" : "Qwen/Qwen2.5-3B-Instruct" ,
163163
164+ // Phi 3 (Microsoft) - Working with fused QKV and RoPE
165+ phi : "microsoft/Phi-3-mini-4k-instruct" ,
166+ phi3 : "microsoft/Phi-3-mini-4k-instruct" ,
167+ "phi-3" : "microsoft/Phi-3-mini-4k-instruct" ,
168+ "phi-3-mini" : "microsoft/Phi-3-mini-4k-instruct" ,
169+
164170 // Llama 3.2 (Meta) - Requires HuggingFace authentication
165171 // Note: meta-llama models require accepting license at huggingface.co
166172 llama : "meta-llama/Llama-3.2-1B-Instruct" ,
167173 "llama-3.2" : "meta-llama/Llama-3.2-1B-Instruct" ,
168174 "llama-3.2-1b" : "meta-llama/Llama-3.2-1B-Instruct" ,
169175 "llama-3.2-3b" : "meta-llama/Llama-3.2-3B-Instruct"
170176
171- // TODO: These models need RoPE fixes in their implementations:
172- // - Phi3/Phi4: Missing RoPE application
173- // - Gemma3n: MoE config parsing + RoPE
174- // - Mistral/Ministral: Missing RoPE application
177+ // TODO: These models need fixes:
178+ // - Gemma3n: Complex AltUp/Laurel architecture
179+ // - Mistral: GQA head count compatibility
175180} as const
176181
177182export type RecommendedModelKey = keyof typeof RECOMMENDED_MODELS
0 commit comments