Skip to content

Commit d24c3ac

Browse files
committed
doc/start: Improve hardware-recommendations.rst
Signed-off-by: Anthony D'Atri <[email protected]>
1 parent 25a6023 commit d24c3ac

File tree

1 file changed

+22
-15
lines changed

1 file changed

+22
-15
lines changed

doc/start/hardware-recommendations.rst

Lines changed: 22 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -68,15 +68,21 @@ is advised.
6868
on a small initial cluster footprint.
6969

7070
There is an :confval:`osd_memory_target` setting for BlueStore OSDs that
71-
defaults to 4GB. Factor in a prudent margin for the operating system and
71+
defaults to 4 GiB. Factor in a prudent margin for the operating system and
7272
administrative tasks (like monitoring and metrics) as well as increased
73-
consumption during recovery: provisioning ~8GB *per BlueStore OSD* is thus
74-
advised.
73+
consumption during recovery. We recommend ensuring that total server RAM
74+
is greater than (number of OSDs * ``osd_memory_target`` * 2), which
75+
allows for usage by the OS and by other Ceph daemons. A 1U server with
76+
8-10 OSDs thus is well-provisioned with 128 GB of physical memory. Enabling
77+
:confval:`osd_memory_target_autotune` can help avoid OOMing under heavy load or when
78+
non-OSD daemons migrate onto a node. An effective :confval:`osd_memory_target` of
79+
at least 6 GiB can help mitigate slow requests on HDD OSDs.
80+
7581

7682
Monitors and Managers (ceph-mon and ceph-mgr)
7783
---------------------------------------------
7884

79-
Monitor and manager daemon memory usage scales with the size of the
85+
Monitor and Manager memory usage scales with the size of the
8086
cluster. Note that at boot-time and during topology changes and recovery these
8187
daemons will need more RAM than they do during steady-state operation, so plan
8288
for peak usage. For very small clusters, 32 GB suffices. For clusters of up to,
@@ -99,8 +105,8 @@ its cache. We recommend 1 GB as a minimum for most systems. See
99105
Memory
100106
======
101107

102-
Bluestore uses its own memory to cache data rather than relying on the
103-
operating system's page cache. In Bluestore you can adjust the amount of memory
108+
BlueStore uses its own memory to cache data rather than relying on the
109+
operating system's page cache. When using the BlueStore OSD back end you can adjust the amount of memory
104110
that the OSD attempts to consume by changing the :confval:`osd_memory_target`
105111
configuration option.
106112

@@ -140,10 +146,11 @@ configuration option.
140146
may result in lower performance, and your Ceph cluster may well be
141147
happier with a daemon that crashes vs one that slows to a crawl.
142148

143-
When using the legacy FileStore back end, the OS page cache was used for caching
144-
data, so tuning was not normally needed. When using the legacy FileStore backend,
145-
the OSD memory consumption was related to the number of PGs per daemon in the
146-
system.
149+
When using the legacy Filestore back end, the OS page cache was used for caching
150+
data, so tuning was not normally needed. OSD memory consumption is related
151+
to the workload and number of PGs that it serves. BlueStore OSDs do not use
152+
the page cache, so the autotuner is recommended to ensure that RAM is used
153+
fully but prudently.
147154

148155

149156
Data Storage
@@ -174,7 +181,7 @@ drives:
174181
For more
175182
information on how to effectively use a mix of fast drives and slow drives in
176183
your Ceph cluster, see the :ref:`block and block.db <bluestore-mixed-device-config>`
177-
section of the Bluestore Configuration Reference.
184+
section of the BlueStore Configuration Reference.
178185

179186
Hard Disk Drives
180187
----------------
@@ -507,19 +514,19 @@ core / spine network switches or routers, often at least 40 Gb/s.
507514
Baseboard Management Controller (BMC)
508515
-------------------------------------
509516

510-
Your server chassis should have a Baseboard Management Controller (BMC).
517+
Your server chassis likely has a Baseboard Management Controller (BMC).
511518
Well-known examples are iDRAC (Dell), CIMC (Cisco UCS), and iLO (HPE).
512519
Administration and deployment tools may also use BMCs extensively, especially
513520
via IPMI or Redfish, so consider the cost/benefit tradeoff of an out-of-band
514-
network for security and administration. Hypervisor SSH access, VM image uploads,
521+
network for security and administration. Hypervisor SSH access, VM image uploads,
515522
OS image installs, management sockets, etc. can impose significant loads on a network.
516523
Running multiple networks may seem like overkill, but each traffic path represents
517524
a potential capacity, throughput and/or performance bottleneck that you should
518525
carefully consider before deploying a large scale data cluster.
519526

520-
Additionally BMCs as of 2023 rarely sport network connections faster than 1 Gb/s,
527+
Additionally, BMCs as of 2025 rarely offer network connections faster than 1 Gb/s,
521528
so dedicated and inexpensive 1 Gb/s switches for BMC administrative traffic
522-
may reduce costs by wasting fewer expenive ports on faster host switches.
529+
may reduce costs by wasting fewer expensive ports on faster host switches.
523530

524531

525532
Failure Domains

0 commit comments

Comments
 (0)