@@ -25,7 +25,6 @@ actor_rollout_ref:
2525 actor :
2626 strategy : fsdp # This is for backward-compatibility
2727 ppo_mini_batch_size : 1536
28- # ppo_micro_batch_size: 8 # will be deprecated, use ppo_micro_batch_size_per_gpu
2928 ppo_micro_batch_size_per_gpu : 1
3029 use_dynamic_bsz : False
3130 ppo_max_token_len_per_gpu : 16384 # n * ${data.max_prompt_length} + ${data.max_response_length}
@@ -57,7 +56,6 @@ actor_rollout_ref:
5756 wrap_policy :
5857 # transformer_layer_cls_to_wrap: None
5958 min_num_params : 0
60- # log_prob_micro_batch_size: 4 # will be deprecated, use log_prob_micro_batch_size_per_gpu
6159 log_prob_micro_batch_size_per_gpu : 1
6260 log_prob_use_dynamic_bsz : ${actor_rollout_ref.actor.use_dynamic_bsz}
6361 log_prob_max_token_len_per_gpu : ${actor_rollout_ref.actor.ppo_max_token_len_per_gpu}
@@ -81,7 +79,6 @@ actor_rollout_ref:
8179 max_num_batched_tokens : 8192
8280 max_model_len : null
8381 max_num_seqs : 1024
84- # log_prob_micro_batch_size: 8 # will be deprecated, use log_prob_micro_batch_size_per_gpu
8582 log_prob_micro_batch_size_per_gpu : 1
8683 log_prob_use_dynamic_bsz : ${actor_rollout_ref.actor.use_dynamic_bsz}
8784 log_prob_max_token_len_per_gpu : ${actor_rollout_ref.actor.ppo_max_token_len_per_gpu}
@@ -115,9 +112,7 @@ critic:
115112 min_num_params : 0
116113 fsdp_size : -1
117114 ppo_mini_batch_size : ${actor_rollout_ref.actor.ppo_mini_batch_size}
118- # ppo_micro_batch_size: 8 # will be deprecated, use ppo_micro_batch_size_per_gpu
119115 ppo_micro_batch_size_per_gpu : 1
120- forward_micro_batch_size : ${critic.ppo_micro_batch_size}
121116 forward_micro_batch_size_per_gpu : ${critic.ppo_micro_batch_size_per_gpu}
122117 use_dynamic_bsz : ${actor_rollout_ref.actor.use_dynamic_bsz}
123118 ppo_max_token_len_per_gpu : 16384 # (${actor_rollout_ref.actor.ppo_max_token_len_per_gpu}) * 2
@@ -140,7 +135,6 @@ reward_model:
140135 min_num_params : 0
141136 param_offload : False
142137 fsdp_size : -1
143- # micro_batch_size: null # will be deprecated, use micro_batch_size_per_gpu
144138 # micro_batch_size_per_gpu: 2 # set a number
145139 # max_length: null
146140 ulysses_sequence_parallel_size : 1 # sp size
0 commit comments