File tree Expand file tree Collapse file tree 2 files changed +1
-6
lines changed Expand file tree Collapse file tree 2 files changed +1
-6
lines changed Original file line number Diff line number Diff line change @@ -412,7 +412,6 @@ void llama_model::load_hparams(llama_model_loader & ml) {
412412    ml.get_key (LLM_KV_BLOCK_COUNT,       hparams.n_layer );
413413    ml.get_key (LLM_KV_EXPERT_COUNT,      hparams.n_expert ,      false );
414414    ml.get_key (LLM_KV_EXPERT_USED_COUNT, hparams.n_expert_used , false );
415-     ml.get_key (LLM_KV_EXPERT_WEIGHTS_SCALE, hparams.expert_weights_scale , false );
416415
417416    if  (arch == LLM_ARCH_WAVTOKENIZER_DEC) {
418417        ml.get_key (LLM_KV_FEATURES_LENGTH, hparams.n_embd_features );
@@ -3702,10 +3701,6 @@ void llama_model::print_info() const {
37023701        LLAMA_LOG_INFO (" %s: f_attention_scale = %f\n "  , __func__, hparams.f_attention_scale );
37033702    }
37043703
3705-     if  (arch == LLM_ARCH_LLAMA) {
3706-         LLAMA_LOG_INFO (" %s: expert_weights_scale = %.1f\n "  , __func__, hparams.expert_weights_scale );
3707-     }
3708- 
37093704    vocab.print_info ();
37103705}
37113706
Original file line number Diff line number Diff line change @@ -1596,7 +1596,7 @@ struct llm_build_context {
15961596                        nullptr ,
15971597                        n_expert, n_expert_used,
15981598                        LLM_FFN_SILU, true ,
1599-                         hparams. expert_weights_scale  >  0 , hparams. expert_weights_scale ,
1599+                         false ,  0.0 ,
16001600                        LLAMA_EXPERT_GATING_FUNC_TYPE_SOFTMAX,
16011601                        cb, il);
16021602                cb (cur, " ffn_moe_out"  , il);
    
 
   
 
     
   
   
          
     
  
    
     
 
    
      
     
 
     
    You can’t perform that action at this time.
  
 
    
  
     
    
      
        
     
 
       
      
     
   
 
    
    
  
 
  
 
     
    
0 commit comments