@@ -8,6 +8,7 @@ TensorFlow Guide to SageMaker's distributed data parallel library
88 - :ref: `tensorflow-sdp-api `
99
1010.. _tensorflow-sdp-modify :
11+ :noindex:
1112
1213Modify a TensorFlow 2.x training script to use SageMaker data parallel
1314======================================================================
@@ -150,6 +151,7 @@ script you will have for distributed training with the library.
150151
151152
152153 .. _tensorflow-sdp-api :
154+ :noindex:
153155
154156TensorFlow API
155157==============
@@ -160,6 +162,7 @@ TensorFlow API
160162
161163
162164.. function :: smdistributed.dataparallel.tensorflow.init()
165+ :noindex:
163166
164167 Initialize ``smdistributed.dataparallel ``. Must be called at the
165168 beginning of the training script.
@@ -183,6 +186,7 @@ TensorFlow API
183186
184187
185188.. function :: smdistributed.dataparallel.tensorflow.size()
189+ :noindex:
186190
187191 The total number of GPUs across all the nodes in the cluster. For
188192 example, in a 8 node cluster with 8 GPUs each, ``size `` will be equal
@@ -200,6 +204,7 @@ TensorFlow API
200204
201205
202206.. function :: smdistributed.dataparallel.tensorflow.local_size()
207+ :noindex:
203208
204209 The total number of GPUs on a node. For example, on a node with 8
205210 GPUs, ``local_size `` will be equal to 8.
@@ -214,6 +219,7 @@ TensorFlow API
214219
215220
216221.. function :: smdistributed.dataparallel.tensorflow.rank()
222+ :noindex:
217223
218224 The rank of the node in the cluster. The rank ranges from 0 to number of
219225 nodes - 1. This is similar to MPI's World Rank.
@@ -228,6 +234,7 @@ TensorFlow API
228234
229235
230236.. function :: smdistributed.dataparallel.tensorflow.local_rank()
237+ :noindex:
231238
232239 Local rank refers to the relative rank of the
233240 GPUs’ ``smdistributed.dataparallel `` processes within the node. For
@@ -246,6 +253,7 @@ TensorFlow API
246253
247254
248255.. function :: smdistributed.dataparallel.tensorflow.allreduce(tensor, param_index, num_params, compression=Compression.none, op=ReduceOp.AVERAGE)
256+ :noindex:
249257
250258 Performs an all-reduce operation on a tensor (``tf.Tensor ``).
251259
@@ -273,6 +281,7 @@ TensorFlow API
273281
274282
275283.. function :: smdistributed.dataparallel.tensorflow.broadcast_global_variables(root_rank)
284+ :noindex:
276285
277286 Broadcasts all global variables from root rank to all other processes.
278287
@@ -287,6 +296,7 @@ TensorFlow API
287296
288297
289298.. function :: smdistributed.dataparallel.tensorflow.broadcast_variables(variables, root_rank)
299+ :noindex:
290300
291301 Applicable for TensorFlow 2.x only.
292302
@@ -309,6 +319,7 @@ TensorFlow API
309319
310320
311321.. function :: smdistributed.dataparallel.tensorflow.oob_allreduce(tensor, compression=Compression.none, op=ReduceOp.AVERAGE)
322+ :noindex:
312323
313324 OutOfBand (oob) AllReduce is simplified AllReduce function for use cases
314325 such as calculating total loss across all the GPUs in the training.
@@ -342,6 +353,7 @@ TensorFlow API
342353
343354
344355.. function :: smdistributed.dataparallel.tensorflow.overlap(tensor)
356+ :noindex:
345357
346358 This function is applicable only for models compiled with XLA. Use this
347359 function to enable ``smdistributed.dataparallel `` to efficiently
@@ -379,6 +391,7 @@ TensorFlow API
379391
380392
381393.. function :: smdistributed.dataparallel.tensorflow.broadcast(tensor, root_rank)
394+ :noindex:
382395
383396 Broadcasts the input tensor on root rank to the same input tensor on all
384397 other ``smdistributed.dataparallel `` processes.
@@ -399,6 +412,7 @@ TensorFlow API
399412
400413
401414.. function:: smdistributed.dataparallel.tensorflow.shutdown()
415+ :noindex:
402416
403417 Shuts down ``smdistributed.dataparallel ``. Optional to call at the end
404418 of the training script.
@@ -413,6 +427,7 @@ TensorFlow API
413427
414428
415429.. function :: smdistributed.dataparallel.tensorflow.DistributedOptimizer
430+ :noindex:
416431
417432 Applicable if you use the ``tf.estimator `` API in TensorFlow 2.x (2.3.1).
418433
@@ -453,6 +468,7 @@ TensorFlow API
453468
454469
455470.. function :: smdistributed.dataparallel.tensorflow.DistributedGradientTape
471+ :noindex:
456472
457473 Applicable to TensorFlow 2.x only.
458474
@@ -488,6 +504,7 @@ TensorFlow API
488504
489505
490506.. function :: smdistributed.dataparallel.tensorflow.BroadcastGlobalVariablesHook
507+ :noindex:
491508
492509 Applicable if you use the ``tf.estimator `` API in TensorFlow 2.x (2.3.1).
493510
@@ -516,6 +533,7 @@ TensorFlow API
516533
517534
518535.. function:: smdistributed.dataparallel.tensorflow.Compression
536+ :noindex:
519537
520538 Optional Gradient Compression algorithm that can be used in AllReduce
521539 operation.
@@ -527,6 +545,7 @@ TensorFlow API
527545
528546
529547.. function :: smdistributed.dataparallel.tensorflow.ReduceOp
548+ :noindex:
530549
531550 Supported reduction operations in ``smdistributed.dataparallel ``.
532551
0 commit comments