@@ -139,12 +139,12 @@ def _send_balancer_worker(self, keys: List[str]) -> str:
139139 existing_keys_set .add (key )
140140 new_keys_added .append (key )
141141
142- if not new_keys_added :
142+ if len ( new_keys_added ) == 0 :
143143 logger .info (f"ℹ️ All { len (keys )} key(s) already exist in balancer" )
144144 return "ok"
145145
146146 # 4. 更新配置中的API_KEYS
147- config_data ['API_KEYS' ] = current_api_keys
147+ config_data ['API_KEYS' ] = existing_keys_set
148148
149149 logger .info (f"📝 Updating gemini balancer config with { len (new_keys_added )} new key(s)..." )
150150
@@ -253,8 +253,6 @@ def _batch_send_worker(self) -> None:
253253 logger .info (f"✅ Gemini balancer queue processed successfully, cleared { len (balancer_keys )} key(s)" )
254254 else :
255255 logger .error (f"❌ Gemini balancer queue processing failed with code: { result_code } " )
256- elif checkpoint .wait_send_balancer and not self .balancer_enabled :
257- logger .info (f"🚫 Gemini Balancer disabled, skipping { len (checkpoint .wait_send_balancer )} key(s) in queue" )
258256
259257 # 发送gpt_load队列
260258 if checkpoint .wait_send_gpt_load and self .gpt_load_enabled :
@@ -269,8 +267,6 @@ def _batch_send_worker(self) -> None:
269267 logger .info (f"✅ GPT load balancer queue processed successfully, cleared { len (gpt_load_keys )} key(s)" )
270268 else :
271269 logger .error (f"❌ GPT load balancer queue processing failed with code: { result_code } " )
272- elif checkpoint .wait_send_gpt_load and not self .gpt_load_enabled :
273- logger .info (f"🚫 GPT Load Balancer disabled, skipping { len (checkpoint .wait_send_gpt_load )} key(s) in queue" )
274270
275271 # 保存checkpoint
276272 self .file_manager .save_checkpoint (checkpoint )
0 commit comments