Skip to content
This repository was archived by the owner on Sep 30, 2022. It is now read-only.

Commit 252cb1e

Browse files
committed
README: minor re-flowing on extra-long lines
No other content changes; just re-flowing of long lines. (cherry picked from commit open-mpi/ompi@e2124c6)
1 parent e059f54 commit 252cb1e

File tree

1 file changed

+33
-30
lines changed

1 file changed

+33
-30
lines changed

README

Lines changed: 33 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -527,37 +527,39 @@ MPI Collectives
527527
MPI process onto Mellanox QDR InfiniBand switch CPUs and HCAs.
528528

529529
- The "ML" coll component is an implementation of MPI collective
530-
operations that takes advantage of communication hierarchies
531-
in modern systems. A ML collective operation is implemented by
530+
operations that takes advantage of communication hierarchies in
531+
modern systems. A ML collective operation is implemented by
532532
combining multiple independently progressing collective primitives
533533
implemented over different communication hierarchies, hence a ML
534-
collective operation is also referred to as a hierarchical collective
535-
operation. The number of collective primitives that are included in a
536-
ML collective operation is a function of subgroups(hierarchies).
537-
Typically, MPI processes in a single communication hierarchy such as
538-
CPU socket, node, or subnet are grouped together into a single subgroup
539-
(hierarchy). The number of subgroups are configurable at runtime,
540-
and each different collective operation could be configured to have
541-
a different of number of subgroups.
534+
collective operation is also referred to as a hierarchical
535+
collective operation. The number of collective primitives that are
536+
included in a ML collective operation is a function of
537+
subgroups(hierarchies). Typically, MPI processes in a single
538+
communication hierarchy such as CPU socket, node, or subnet are
539+
grouped together into a single subgroup (hierarchy). The number of
540+
subgroups are configurable at runtime, and each different collective
541+
operation could be configured to have a different of number of
542+
subgroups.
542543

543544
The component frameworks and components used by/required for a
544545
"ML" collective operation.
545546

546547
Frameworks:
547-
* "sbgp" - Provides functionality for grouping processes into subgroups
548+
* "sbgp" - Provides functionality for grouping processes into
549+
subgroups
548550
* "bcol" - Provides collective primitives optimized for a particular
549551
communication hierarchy
550552

551553
Components:
552-
* sbgp components - Provides grouping functionality over a CPU socket
553-
("basesocket"), shared memory ("basesmuma"),
554-
Mellanox's ConnectX HCA ("ibnet"), and other
555-
interconnects supported by PML ("p2p")
556-
557-
* BCOL components - Provides optimized collective primitives for
558-
shared memory ("basesmuma"), Mellanox's ConnectX
559-
HCA ("iboffload"), and other interconnects supported
560-
by PML ("ptpcoll")
554+
* sbgp components - Provides grouping functionality over a CPU
555+
socket ("basesocket"), shared memory
556+
("basesmuma"), Mellanox's ConnectX HCA
557+
("ibnet"), and other interconnects supported by
558+
PML ("p2p")
559+
* BCOL components - Provides optimized collective primitives for
560+
shared memory ("basesmuma"), Mellanox's ConnectX
561+
HCA ("iboffload"), and other interconnects
562+
supported by PML ("ptpcoll")
561563

562564
- The "cuda" coll component provides CUDA-aware support for the
563565
reduction type collectives with GPU buffers. This component is only
@@ -997,10 +999,11 @@ RUN-TIME SYSTEM SUPPORT
997999
most cases. This option is only needed for special configurations.
9981000

9991001
--with-pmi
1000-
Build PMI support (by default on non-Cray XE/XC systems, it is not built).
1001-
On Cray XE/XC systems, the location of pmi is detected automatically as
1002-
part of the configure process. For non-Cray systems, if the pmi2.h header
1003-
is found in addition to pmi.h, then support for PMI2 will be built.
1002+
Build PMI support (by default on non-Cray XE/XC systems, it is not
1003+
built). On Cray XE/XC systems, the location of pmi is detected
1004+
automatically as part of the configure process. For non-Cray
1005+
systems, if the pmi2.h header is found in addition to pmi.h, then
1006+
support for PMI2 will be built.
10041007

10051008
--with-slurm
10061009
Force the building of SLURM scheduler support.
@@ -1631,9 +1634,9 @@ Open MPI API Extensions
16311634
-----------------------
16321635

16331636
Open MPI contains a framework for extending the MPI API that is
1634-
available to applications. Each extension is usually a standalone set of
1635-
functionality that is distinct from other extensions (similar to how
1636-
Open MPI's plugins are usually unrelated to each other). These
1637+
available to applications. Each extension is usually a standalone set
1638+
of functionality that is distinct from other extensions (similar to
1639+
how Open MPI's plugins are usually unrelated to each other). These
16371640
extensions provide new functions and/or constants that are available
16381641
to MPI applications.
16391642

@@ -1951,9 +1954,9 @@ Here's how the three sub-groups are defined:
19511954
get their MPI/OSHMEM application to run correctly.
19521955
2. Application tuner: Generally, these are parameters that can be
19531956
used to tweak MPI application performance.
1954-
3. MPI/OSHMEM developer: Parameters that either don't fit in the other two,
1955-
or are specifically intended for debugging / development of Open
1956-
MPI itself.
1957+
3. MPI/OSHMEM developer: Parameters that either don't fit in the
1958+
other two, or are specifically intended for debugging /
1959+
development of Open MPI itself.
19571960

19581961
Each sub-group is broken down into three classifications:
19591962

0 commit comments

Comments
 (0)