@@ -68,15 +68,21 @@ is advised.
6868 on a small initial cluster footprint.
6969
7070There is an :confval: `osd_memory_target ` setting for BlueStore OSDs that
71- defaults to 4GB. Factor in a prudent margin for the operating system and
71+ defaults to 4 GiB. Factor in a prudent margin for the operating system and
7272administrative tasks (like monitoring and metrics) as well as increased
73- consumption during recovery: provisioning ~8GB *per BlueStore OSD * is thus
74- advised.
73+ consumption during recovery. We recommend ensuring that total server RAM
74+ is greater than (number of OSDs * ``osd_memory_target `` * 2), which
75+ allows for usage by the OS and by other Ceph daemons. A 1U server with
76+ 8-10 OSDs thus is well-provisioned with 128 GB of physical memory. Enabling
77+ :confval: `osd_memory_target_autotune ` can help avoid OOMing under heavy load or when
78+ non-OSD daemons migrate onto a node. An effective :confval: `osd_memory_target ` of
79+ at least 6 GiB can help mitigate slow requests on HDD OSDs.
80+
7581
7682Monitors and Managers (ceph-mon and ceph-mgr)
7783---------------------------------------------
7884
79- Monitor and manager daemon memory usage scales with the size of the
85+ Monitor and Manager memory usage scales with the size of the
8086cluster. Note that at boot-time and during topology changes and recovery these
8187daemons will need more RAM than they do during steady-state operation, so plan
8288for peak usage. For very small clusters, 32 GB suffices. For clusters of up to,
@@ -99,8 +105,8 @@ its cache. We recommend 1 GB as a minimum for most systems. See
99105Memory
100106======
101107
102- Bluestore uses its own memory to cache data rather than relying on the
103- operating system's page cache. In Bluestore you can adjust the amount of memory
108+ BlueStore uses its own memory to cache data rather than relying on the
109+ operating system's page cache. When using the BlueStore OSD back end you can adjust the amount of memory
104110that the OSD attempts to consume by changing the :confval: `osd_memory_target `
105111configuration option.
106112
@@ -140,10 +146,11 @@ configuration option.
140146 may result in lower performance, and your Ceph cluster may well be
141147 happier with a daemon that crashes vs one that slows to a crawl.
142148
143- When using the legacy FileStore back end, the OS page cache was used for caching
144- data, so tuning was not normally needed. When using the legacy FileStore backend,
145- the OSD memory consumption was related to the number of PGs per daemon in the
146- system.
149+ When using the legacy Filestore back end, the OS page cache was used for caching
150+ data, so tuning was not normally needed. OSD memory consumption is related
151+ to the workload and number of PGs that it serves. BlueStore OSDs do not use
152+ the page cache, so the autotuner is recommended to ensure that RAM is used
153+ fully but prudently.
147154
148155
149156Data Storage
@@ -174,7 +181,7 @@ drives:
174181For more
175182information on how to effectively use a mix of fast drives and slow drives in
176183your Ceph cluster, see the :ref: `block and block.db <bluestore-mixed-device-config >`
177- section of the Bluestore Configuration Reference.
184+ section of the BlueStore Configuration Reference.
178185
179186Hard Disk Drives
180187----------------
@@ -507,19 +514,19 @@ core / spine network switches or routers, often at least 40 Gb/s.
507514Baseboard Management Controller (BMC)
508515-------------------------------------
509516
510- Your server chassis should have a Baseboard Management Controller (BMC).
517+ Your server chassis likely has a Baseboard Management Controller (BMC).
511518Well-known examples are iDRAC (Dell), CIMC (Cisco UCS), and iLO (HPE).
512519Administration and deployment tools may also use BMCs extensively, especially
513520via IPMI or Redfish, so consider the cost/benefit tradeoff of an out-of-band
514- network for security and administration. Hypervisor SSH access, VM image uploads,
521+ network for security and administration. Hypervisor SSH access, VM image uploads,
515522OS image installs, management sockets, etc. can impose significant loads on a network.
516523Running multiple networks may seem like overkill, but each traffic path represents
517524a potential capacity, throughput and/or performance bottleneck that you should
518525carefully consider before deploying a large scale data cluster.
519526
520- Additionally BMCs as of 2023 rarely sport network connections faster than 1 Gb/s,
527+ Additionally, BMCs as of 2025 rarely offer network connections faster than 1 Gb/s,
521528so dedicated and inexpensive 1 Gb/s switches for BMC administrative traffic
522- may reduce costs by wasting fewer expenive ports on faster host switches.
529+ may reduce costs by wasting fewer expensive ports on faster host switches.
523530
524531
525532Failure Domains
0 commit comments