Skip to content

Commit 859fbf0

Browse files
committed
Fix errors and add to index
1 parent 50309eb commit 859fbf0

8 files changed

+83
-70
lines changed

doc/source/operations/baremetal-node-management.rst

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -181,7 +181,7 @@ The command below extracts a list of port UUID, node UUID and switch port inform
181181

182182
.. code-block:: bash
183183
184-
admin# openstack baremetal port list --field uuid --field node_uuid --field local_link_connection --format value
184+
openstack baremetal port list --field uuid --field node_uuid --field local_link_connection --format value
185185
186186
NGS will manage VLAN membership for ports when the ``local_link_connection`` fields match one of the switches in ``ml2_conf.ini``.
187187
The rest of the switch configuration is static.

doc/source/operations/ceph-management.rst

Lines changed: 34 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -45,8 +45,8 @@ Ceph commands are usually run inside a ``cephadm shell`` utility container:
4545

4646
.. code-block:: console
4747
48-
# From the node that runs Ceph
49-
ceph# sudo cephadm shell
48+
# From storage host
49+
sudo cephadm shell
5050
5151
Operating a cluster requires a keyring with an admin access to be available for Ceph
5252
commands. Cephadm will copy such keyring to the nodes carrying
@@ -71,15 +71,17 @@ First drain the node
7171

7272
.. code-block:: console
7373
74-
ceph# cephadm shell
75-
ceph# ceph orch host drain <host>
74+
# From storage host
75+
sudo cephadm shell
76+
ceph orch host drain <host>
7677
7778
Once all daemons are removed - you can remove the host:
7879

7980
.. code-block:: console
8081
81-
ceph# cephadm shell
82-
ceph# ceph orch host rm <host>
82+
# From storage host
83+
sudo cephadm shell
84+
ceph orch host rm <host>
8385
8486
And then remove the host from inventory (usually in
8587
``etc/kayobe/inventory/overcloud``)
@@ -98,24 +100,27 @@ movement:
98100

99101
.. code-block:: console
100102
101-
ceph# cephadm shell
102-
ceph# ceph osd set noout
103+
# From storage host
104+
sudo cephadm shell
105+
ceph osd set noout
103106
104107
Reboot the node and replace the drive
105108

106109
Unset noout after the node is back online
107110

108111
.. code-block:: console
109112
110-
ceph# cephadm shell
111-
ceph# ceph osd unset noout
113+
# From storage host
114+
sudo cephadm shell
115+
ceph osd unset noout
112116
113117
Remove the OSD using Ceph orchestrator command:
114118

115119
.. code-block:: console
116120
117-
ceph# cephadm shell
118-
ceph# ceph orch osd rm <ID> --replace
121+
# From storage host
122+
sudo cephadm shell
123+
ceph orch osd rm <ID> --replace
119124
120125
After removing OSDs, if the drives the OSDs were deployed on once again become
121126
available, cephadm may automatically try to deploy more OSDs on these drives if
@@ -142,7 +147,7 @@ identify which OSDs are tied to which physical disks:
142147

143148
.. code-block:: console
144149
145-
ceph# ceph device ls
150+
ceph device ls
146151
147152
Host maintenance
148153
----------------
@@ -167,7 +172,7 @@ Ceph can report details about failed OSDs by running:
167172

168173
.. code-block:: console
169174
170-
ceph# ceph health detail
175+
ceph health detail
171176
172177
.. note ::
173178
@@ -184,15 +189,16 @@ A failed OSD will also be reported as down by running:
184189

185190
.. code-block:: console
186191
187-
ceph# ceph osd tree
192+
ceph osd tree
188193
189194
Note the ID of the failed OSD.
190195

191196
The failed disk is usually logged by the Linux kernel too:
192197

193198
.. code-block:: console
194199
195-
storage-0# dmesg -T
200+
# From storage host
201+
dmesg -T
196202
197203
Cross-reference the hardware device and OSD ID to ensure they match.
198204
(Using `pvs` and `lvs` may help make this connection).
@@ -207,36 +213,35 @@ show``).
207213
On this hypervisor, enter the libvirt container:
208214

209215
.. code-block:: console
210-
:substitutions:
211216
212-
|hypervisor_hostname|# docker exec -it nova_libvirt /bin/bash
217+
# From hypervisor host
218+
docker exec -it nova_libvirt /bin/bash
213219
214220
Find the VM name using libvirt:
215221

216222
.. code-block:: console
217-
:substitutions:
218223
219-
(nova-libvirt)[root@|hypervisor_hostname| /]# virsh list
224+
(nova-libvirt)[root@compute-01 /]# virsh list
220225
Id Name State
221226
------------------------------------
222227
1 instance-00000001 running
223228
224229
Now inspect the properties of the VM using ``virsh dumpxml``:
225230

226231
.. code-block:: console
227-
:substitutions:
228232
229-
(nova-libvirt)[root@|hypervisor_hostname| /]# virsh dumpxml instance-00000001 | grep rbd
230-
<source protocol='rbd' name='|nova_rbd_pool|/51206278-e797-4153-b720-8255381228da_disk'>
233+
(nova-libvirt)[root@compute-01 /]# virsh dumpxml instance-00000001 | grep rbd
234+
<source protocol='rbd' name='<nova rbd pool>/51206278-e797-4153-b720-8255381228da_disk'>
231235
232236
On a Ceph node, the RBD pool can be inspected and the volume extracted as a RAW
233237
block image:
234238

235239
.. code-block:: console
236-
:substitutions:
237240
238-
ceph# rbd ls |nova_rbd_pool|
239-
ceph# rbd export |nova_rbd_pool|/51206278-e797-4153-b720-8255381228da_disk blob.raw
241+
# From storage host
242+
sudo cephadm shell
243+
rbd ls <nova rbd pool>
244+
rbd export <nova rbd pool>/51206278-e797-4153-b720-8255381228da_disk blob.raw
240245
241246
The raw block device (blob.raw above) can be mounted using the loopback device.
242247

@@ -248,8 +253,9 @@ libguestfs-tools and using the guestfish command:
248253

249254
.. code-block:: console
250255
251-
ceph# export LIBGUESTFS_BACKEND=direct
252-
ceph# guestfish -a blob.qcow
256+
# From storage host
257+
export LIBGUESTFS_BACKEND=direct
258+
guestfish -a blob.qcow
253259
><fs> run
254260
100% [XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX] 00:00
255261
><fs> list-filesystems

doc/source/operations/control-plane-operation.rst

Lines changed: 24 additions & 23 deletions
Original file line numberDiff line numberDiff line change
@@ -55,7 +55,7 @@ Configuring Prometheus Alerts
5555
-----------------------------
5656

5757
Alerts are defined in code and stored in Kayobe configuration. See ``*.rules``
58-
files in ``${KAYOBE_CONFIG_PATH}/kolla/config/prometheus`` as a model to add
58+
files in ``$KAYOBE_CONFIG_PATH/kolla/config/prometheus`` as a model to add
5959
custom rules.
6060

6161
Silencing Prometheus Alerts
@@ -88,7 +88,7 @@ Generating Alerts from Metrics
8888
++++++++++++++++++++++++++++++
8989

9090
Alerts are defined in code and stored in Kayobe configuration. See ``*.rules``
91-
files in ``${KAYOBE_CONFIG_PATH}/kolla/config/prometheus`` as a model to add
91+
files in ``$KAYOBE_CONFIG_PATH/kolla/config/prometheus`` as a model to add
9292
custom rules.
9393

9494
Control Plane Shutdown Procedure
@@ -124,7 +124,7 @@ The password can be found using:
124124

125125
.. code-block:: console
126126
127-
kayobe# ansible-vault view ${KAYOBE_CONFIG_PATH}/kolla/passwords.yml \
127+
kayobe# ansible-vault view $KAYOBE_CONFIG_PATH/kolla/passwords.yml \
128128
--vault-password-file <Vault password file path> | grep ^database
129129
130130
Checking RabbitMQ
@@ -135,6 +135,7 @@ RabbitMQ health is determined using the command ``rabbitmqctl cluster_status``:
135135
.. code-block:: console
136136
137137
[stack@controller0 ~]$ docker exec rabbitmq rabbitmqctl cluster_status
138+
138139
Cluster status of node rabbit@controller0 ...
139140
[{nodes,[{disc,['rabbit@controller0','rabbit@controller1',
140141
'rabbit@controller2']}]},
@@ -180,20 +181,18 @@ If you are shutting down a single hypervisor, to avoid down time to tenants it
180181
is advisable to migrate all of the instances to another machine. See
181182
:ref:`evacuating-all-instances`.
182183

183-
.. ifconfig:: deployment['ceph_managed']
184-
185-
Ceph
186-
----
184+
Ceph
185+
----
187186

188-
The following guide provides a good overview:
189-
https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/8/html/director_installation_and_usage/sect-rebooting-ceph
187+
The following guide provides a good overview:
188+
https://access.redhat.com/documentation/en-us/red_hat_openstack_platform/8/html/director_installation_and_usage/sect-rebooting-ceph
190189

191190
Shutting down the seed VM
192191
-------------------------
193192

194193
.. code-block:: console
195194
196-
kayobe# virsh shutdown <Seed node>
195+
kayobe# virsh shutdown <Seed hostname>
197196
198197
.. _full-shutdown:
199198

@@ -262,7 +261,7 @@ hypervisor is powered on. If it does not, it can be started with:
262261

263262
.. code-block:: console
264263
265-
kayobe# virsh start seed-0
264+
kayobe# virsh start <Seed hostname>
266265
267266
Full power on
268267
-------------
@@ -340,13 +339,14 @@ To see the list of hypervisor names:
340339

341340
.. code-block:: console
342341
343-
admin# openstack hypervisor list
342+
# From host that can reach Openstack
343+
openstack hypervisor list
344344
345345
To boot an instance on a specific hypervisor
346346

347347
.. code-block:: console
348348
349-
admin# openstack server create --flavor <Flavour name>--network <Network name> --key-name <key> --image <Image name> --availability-zone nova::<Hypervisor name> <VM name>
349+
openstack server create --flavor <Flavour name>--network <Network name> --key-name <key> --image <Image name> --availability-zone nova::<Hypervisor name> <VM name>
350350
351351
Cleanup Procedures
352352
==================
@@ -360,22 +360,23 @@ perform the following cleanup procedure regularly:
360360

361361
.. code-block:: console
362362
363-
admin# for user in $(openstack user list --domain magnum -f value -c Name | grep -v magnum_trustee_domain_admin); do
364-
if openstack coe cluster list -c uuid -f value | grep -q $(echo $user | sed 's/_[0-9a-f]*$//'); then
365-
echo "$user still in use, not deleting"
366-
else
367-
openstack user delete --domain magnum $user
368-
fi
369-
done
363+
for user in $(openstack user list --domain magnum -f value -c Name | grep -v magnum_trustee_domain_admin); do
364+
if openstack coe cluster list -c uuid -f value | grep -q $(echo $user | sed 's/_[0-9a-f]*$//'); then
365+
echo "$user still in use, not deleting"
366+
else
367+
openstack user delete --domain magnum $user
368+
fi
369+
done
370370
371371
OpenSearch indexes retention
372372
=============================
373373

374374
To alter default rotation values for OpenSearch, edit
375375

376-
``${KAYOBE_CONFIG_PATH}/kolla/globals.yml``:
376+
``$KAYOBE_CONFIG_PATH/kolla/globals.yml``:
377377

378378
.. code-block:: console
379+
379380
# Duration after which index is closed (default 30)
380381
opensearch_soft_retention_period_days: 90
381382
# Duration after which index is deleted (default 60)
@@ -384,8 +385,8 @@ To alter default rotation values for OpenSearch, edit
384385
Reconfigure Opensearch with new values:
385386

386387
.. code-block:: console
387-
kayobe overcloud service reconfigure --kolla-tags opensearch
388388
389-
For more information see the `upstream documentation
389+
kayobe# kayobe overcloud service reconfigure --kolla-tags opensearch
390390
391+
For more information see the `upstream documentation
391392
<https://docs.openstack.org/kolla-ansible/latest/reference/logging-and-monitoring/central-logging-guide.html#applying-log-retention-policies>`__.

doc/source/operations/customising_horizon.rst renamed to doc/source/operations/customising-horizon.rst

Lines changed: 2 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,8 +1,6 @@
1-
.. include:: vars.rst
2-
3-
====================================
1+
===================
42
Customising Horizon
5-
====================================
3+
===================
64

75
Horizon is the most frequent site-specific container customisation required:
86
other customisations tend to be common across deployments, but personalisation
@@ -55,7 +53,6 @@ Building a custom container image for Horizon can be done by modifying
5553
``kolla.yml`` to fetch the custom theme and include it in the image:
5654

5755
.. code-block:: yaml
58-
:substitutions:
5956
6057
kolla_sources:
6158
horizon-additions-theme-<custom theme name>:

doc/source/operations/hardware-inventory-management.rst

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ Hardware Inventory Management
55
At its lowest level, hardware inventory is managed in the Bifrost service.
66

77
Reconfiguring Control Plane Hardware
8-
------------------------------------
8+
====================================
99

1010
If a server's hardware or firmware configuration is changed, it should be
1111
re-inspected in Bifrost before it is redeployed into service. A single server
@@ -112,10 +112,10 @@ hypervisor. They should all show the status ACTIVE. This can be verified with:
112112
admin# openstack server show <instance uuid>
113113
114114
Troubleshooting
115-
+++++++++++++++
115+
===============
116116

117117
Servers that have been shut down
118-
********************************
118+
--------------------------------
119119

120120
If there are any instances that are SHUTOFF they won’t be migrated, but you can
121121
use ``openstack server migrate`` for them once the live migration is finished.
@@ -131,7 +131,7 @@ For more details see:
131131
http://www.danplanet.com/blog/2016/03/03/evacuate-in-nova-one-command-to-confuse-us-all/
132132

133133
Flavors have changed
134-
********************
134+
--------------------
135135

136136
If the size of the flavors has changed, some instances will also fail to
137137
migrate as the process needs manual confirmation. You can do this with:
@@ -150,7 +150,7 @@ RESIZE`` as shown in this snippet of ``openstack server show <instance-uuid>``:
150150
.. _set-bifrost-maintenance-mode:
151151

152152
Set maintenance mode on a node in Bifrost
153-
+++++++++++++++++++++++++++++++++++++++++
153+
-----------------------------------------
154154

155155
.. code-block:: console
156156
@@ -161,7 +161,7 @@ Set maintenance mode on a node in Bifrost
161161
.. _unset-bifrost-maintenance-mode:
162162

163163
Unset maintenance mode on a node in Bifrost
164-
+++++++++++++++++++++++++++++++++++++++++++
164+
-------------------------------------------
165165

166166
.. code-block:: console
167167

doc/source/operations/index.rst

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,10 +7,20 @@ This guide is for operators of the StackHPC Kayobe configuration project.
77
.. toctree::
88
:maxdepth: 1
99

10+
baremetal-node-management
11+
ceph-management
12+
control-plane-operation
13+
customsing-horizon
14+
gpu-in-openstack
15+
hardware-inventory-management
1016
hotfix-playbook
17+
migrating-vm
1118
nova-compute-ironic
1219
octavia
20+
openstack-projects-and-users-management
21+
openstack-reconfiguration
1322
rabbitmq
1423
secret-rotation
1524
tempest
1625
upgrading
26+
wazuh-operation

0 commit comments

Comments
 (0)