@@ -84,7 +84,7 @@ def supervised_training_step(
8484 trainer = Engine(update_fn)
8585
8686 .. versionadded:: 0.4.5
87- .. versionchanged:: 0.5.0
87+ .. versionchanged:: 0.4.7
8888 Added Gradient Accumulation.
8989 """
9090
@@ -158,7 +158,7 @@ def supervised_training_step_amp(
158158 trainer = Engine(update_fn)
159159
160160 .. versionadded:: 0.4.5
161- .. versionchanged:: 0.5.0
161+ .. versionchanged:: 0.4.7
162162 Added Gradient Accumulation.
163163 """
164164
@@ -241,7 +241,7 @@ def supervised_training_step_apex(
241241 trainer = Engine(update_fn)
242242
243243 .. versionadded:: 0.4.5
244- .. versionchanged:: 0.5.0
244+ .. versionchanged:: 0.4.7
245245 Added Gradient Accumulation.
246246 """
247247
@@ -318,7 +318,7 @@ def supervised_training_step_tpu(
318318 trainer = Engine(update_fn)
319319
320320 .. versionadded:: 0.4.5
321- .. versionchanged:: 0.5.0
321+ .. versionchanged:: 0.4.7
322322 Added Gradient Accumulation argument for all supervised training methods.
323323 """
324324 try :
@@ -494,7 +494,7 @@ def output_transform_fn(x, y, y_pred, loss):
494494 - Added ``amp_mode`` argument for automatic mixed precision.
495495 - Added ``scaler`` argument for gradient scaling.
496496
497- .. versionchanged:: 0.5.0
497+ .. versionchanged:: 0.4.7
498498 Added Gradient Accumulation argument for all supervised training methods.
499499 """
500500
0 commit comments