Whether can use 4 GPU and 1/4 stop_batchs to reduce training time and achieve same accuracy. #4949
-
Dear Maintainers, |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 3 replies
-
You may increase the number of fp data collected in each iteration to reduce the total number of iterations, thus reducing the total training cost. |
Beta Was this translation helpful? Give feedback.
using 4 GPU effective increase the batch size by 4 times, but it does not always leads to faster decay of the error by 4 times. One may have to test case by case
for dpa-1 and se_a descriptors, smaller batch size (like auto:32) is usually preferred, while for dpa-2 and dpa-3, larger batch size may speedup the training.