Skip to content

Commit 15935db

Browse files
committed
doc: Use existing labels and ref for hyperlinks in architecture.rst
Use validated ":ref:" hyperlinks instead of "external links" in "target definitions" when linking within the Ceph docs: - Update to use existing labels when linkin from architecture.rst. - Remove unused "target definitions". Also use title case for section titles in doc/start/hardware-recommendations.rst because change to use link text generated from section title. Other than generated link texts the rendered PR should look the same as the old docs, only differing in the source RST. Signed-off-by: Ville Ojamo <[email protected]>
1 parent b524951 commit 15935db

File tree

2 files changed

+14
-25
lines changed

2 files changed

+14
-25
lines changed

doc/architecture.rst

Lines changed: 11 additions & 22 deletions
Original file line numberDiff line numberDiff line change
@@ -206,7 +206,7 @@ monitor, two in a cluster that contains three monitors, three in a cluster that
206206
contains five monitors, four in a cluster that contains six monitors, and so
207207
on).
208208

209-
See the `Monitor Config Reference`_ for more detail on configuring monitors.
209+
See the :ref:`monitor-config-reference` for more detail on configuring monitors.
210210

211211
.. index:: architecture; high availability authentication
212212

@@ -368,9 +368,9 @@ daemons. The authentication is not extended beyond the Ceph client. If a user
368368
accesses the Ceph client from a remote host, cephx authentication will not be
369369
applied to the connection between the user's host and the client host.
370370

371-
See `Cephx Config Guide`_ for more on configuration details.
371+
See :ref:`rados-cephx-config-ref` for more on configuration details.
372372

373-
See `User Management`_ for more on user management.
373+
See :ref:`user-management` for more on user management.
374374

375375
See :ref:`A Detailed Description of the Cephx Authentication Protocol
376376
<cephx_2012_peter>` for more on the distinction between authorization and
@@ -433,8 +433,8 @@ the greater cluster provides several benefits:
433433
mismatches in object size and finds metadata mismatches, and is usually
434434
performed daily. Ceph OSD Daemons perform deeper scrubbing by comparing the
435435
data in objects, bit-for-bit, against their checksums. Deep scrubbing finds
436-
bad sectors on drives that are not detectable with light scrubs. See `Data
437-
Scrubbing`_ for details on configuring scrubbing.
436+
bad sectors on drives that are not detectable with light scrubs. See :ref:`Data
437+
Scrubbing <rados_config_scrubbing>` for details on configuring scrubbing.
438438

439439
#. **Replication:** Data replication involves collaboration between Ceph
440440
Clients and Ceph OSD Daemons. Ceph OSD Daemons use the CRUSH algorithm to
@@ -525,7 +525,7 @@ Pools set at least the following parameters:
525525
- The Number of Placement Groups, and
526526
- The CRUSH Rule to Use.
527527

528-
See `Set Pool Values`_ for details.
528+
See :ref:`setpoolvalues` for details.
529529

530530

531531
.. index: architecture; placement group mapping
@@ -626,7 +626,7 @@ which is the process of bringing all of the OSDs that store a Placement Group
626626
(PG) into agreement about the state of all of the RADOS objects (and their
627627
metadata) in that PG. Ceph OSD Daemons `Report Peering Failure`_ to the Ceph
628628
Monitors. Peering issues usually resolve themselves; however, if the problem
629-
persists, you may need to refer to the `Troubleshooting Peering Failure`_
629+
persists, you may need to refer to the :ref:`Troubleshooting Peering Failure <failures-osd-peering>`
630630
section.
631631

632632
.. Note:: PGs that agree on the state of the cluster do not necessarily have
@@ -721,7 +721,7 @@ scrubbing by comparing data in objects bit-for-bit. Deep scrubbing (by default
721721
performed weekly) finds bad blocks on a drive that weren't apparent in a light
722722
scrub.
723723

724-
See `Data Scrubbing`_ for details on configuring scrubbing.
724+
See :ref:`Data Scrubbing <rados_config_scrubbing>` for details on configuring scrubbing.
725725

726726

727727

@@ -1219,8 +1219,8 @@ appliances do not fully utilize the CPU and RAM of a typical commodity server,
12191219
Ceph does. From heartbeats, to peering, to rebalancing the cluster or
12201220
recovering from faults, Ceph offloads work from clients (and from a centralized
12211221
gateway which doesn't exist in the Ceph architecture) and uses the computing
1222-
power of the OSDs to perform the work. When referring to `Hardware
1223-
Recommendations`_ and the `Network Config Reference`_, be cognizant of the
1222+
power of the OSDs to perform the work. When referring to :ref:`hardware-recommendations`
1223+
and the `Network Config Reference`_, be cognizant of the
12241224
foregoing concepts to understand how Ceph utilizes computing resources.
12251225

12261226
.. index:: Ceph Protocol, librados
@@ -1574,7 +1574,7 @@ another application.
15741574
correspond in a 1:1 manner with an object stored in the storage cluster. It
15751575
is possible for an S3 or Swift object to map to multiple Ceph objects.
15761576

1577-
See `Ceph Object Storage`_ for details.
1577+
See :ref:`object-gateway` for details.
15781578

15791579

15801580
.. index:: Ceph Block Device; block device; RBD; Rados Block Device
@@ -1671,26 +1671,15 @@ instance for high availability.
16711671

16721672
.. _RADOS - A Scalable, Reliable Storage Service for Petabyte-scale Storage Clusters: https://ceph.io/assets/pdfs/weil-rados-pdsw07.pdf
16731673
.. _Paxos: https://en.wikipedia.org/wiki/Paxos_(computer_science)
1674-
.. _Monitor Config Reference: ../rados/configuration/mon-config-ref
1675-
.. _Monitoring OSDs and PGs: ../rados/operations/monitoring-osd-pg
16761674
.. _Heartbeats: ../rados/configuration/mon-osd-interaction
16771675
.. _Monitoring OSDs: ../rados/operations/monitoring-osd-pg/#monitoring-osds
16781676
.. _CRUSH - Controlled, Scalable, Decentralized Placement of Replicated Data: https://ceph.io/assets/pdfs/weil-crush-sc06.pdf
1679-
.. _Data Scrubbing: ../rados/configuration/osd-config-ref#scrubbing
16801677
.. _Report Peering Failure: ../rados/configuration/mon-osd-interaction#osds-report-peering-failure
1681-
.. _Troubleshooting Peering Failure: ../rados/troubleshooting/troubleshooting-pg#placement-group-down-peering-failure
1682-
.. _Ceph Authentication and Authorization: ../rados/operations/auth-intro/
1683-
.. _Hardware Recommendations: ../start/hardware-recommendations
16841678
.. _Network Config Reference: ../rados/configuration/network-config-ref
1685-
.. _Data Scrubbing: ../rados/configuration/osd-config-ref#scrubbing
16861679
.. _striping: https://en.wikipedia.org/wiki/Data_striping
16871680
.. _RAID: https://en.wikipedia.org/wiki/RAID
16881681
.. _RAID 0: https://en.wikipedia.org/wiki/RAID_0#RAID_0
1689-
.. _Ceph Object Storage: ../radosgw/
16901682
.. _RESTful: https://en.wikipedia.org/wiki/RESTful
16911683
.. _Erasure Code Notes: https://github.com/ceph/ceph/blob/40059e12af88267d0da67d8fd8d9cd81244d8f93/doc/dev/osd_internals/erasure_coding/developer_notes.rst
16921684
.. _Cache Tiering: ../rados/operations/cache-tiering
1693-
.. _Set Pool Values: ../rados/operations/pools#set-pool-values
16941685
.. _Kerberos: https://en.wikipedia.org/wiki/Kerberos_(protocol)
1695-
.. _Cephx Config Guide: ../rados/configuration/auth-config-ref
1696-
.. _User Management: ../rados/operations/user-management

doc/start/hardware-recommendations.rst

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
.. _hardware-recommendations:
22

33
==========================
4-
hardware recommendations
4+
Hardware Recommendations
55
==========================
66

77
Ceph is designed to run on commodity hardware, which makes building and
@@ -73,7 +73,7 @@ administrative tasks (like monitoring and metrics) as well as increased
7373
consumption during recovery: provisioning ~8GB *per BlueStore OSD* is thus
7474
advised.
7575

76-
Monitors and managers (ceph-mon and ceph-mgr)
76+
Monitors and Managers (ceph-mon and ceph-mgr)
7777
---------------------------------------------
7878

7979
Monitor and manager daemon memory usage scales with the size of the
@@ -88,7 +88,7 @@ tuning the following settings:
8888
* :confval:`rocksdb_cache_size`
8989

9090

91-
Metadata servers (ceph-mds)
91+
Metadata Servers (ceph-mds)
9292
---------------------------
9393

9494
CephFS metadata daemon memory utilization depends on the configured size of

0 commit comments

Comments
 (0)