Skip to content

Commit 37813cf

Browse files
authored
Merge pull request ceph#65721 from phlogistonjohn/jjm-qa-task-template
qa: Add new task module template.py Reviewed-by: Adam King <[email protected]>
2 parents 85e6b63 + 30f8299 commit 37813cf

30 files changed

+221
-227
lines changed

qa/suites/nvmeof/thrash/gateway-initiator-setup/10-subsys-90-namespace-no_huge_pages.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ tasks:
1313
- cephadm.wait_for_service:
1414
service: nvmeof.mypool.mygroup0
1515

16-
- cephadm.exec:
16+
- exec:
1717
host.a:
1818
- ceph orch ls nvmeof --export > /tmp/nvmeof-orig.yaml
1919
- cp /tmp/nvmeof-orig.yaml /tmp/nvmeof-no-huge-page.yaml

qa/suites/orch/cephadm/mgr-nfs-upgrade/2-nfs.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
tasks:
22

33
# stop kernel nfs server, if running
4-
- vip.exec:
4+
- exec:
55
all-hosts:
66
- systemctl stop nfs-server
77

@@ -20,7 +20,7 @@ tasks:
2020
# we can't do wait_for_service here because with octopus it's nfs.ganesha-foo not nfs.foo
2121
- while ! ceph orch ls | grep nfs | grep 2/2 ; do sleep 1 ; done
2222

23-
- vip.exec:
23+
- exec:
2424
host.a:
2525
- mkdir /mnt/foo
2626
- while ! mount -t nfs $(hostname):/fake /mnt/foo -o sync ; do sleep 5 ; done

qa/suites/orch/cephadm/mgr-nfs-upgrade/4-final.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -3,7 +3,7 @@ overrides:
33
log-ignorelist:
44
- CEPHADM_REFRESH_FAILED
55
tasks:
6-
- vip.exec:
6+
- exec:
77
host.a:
88
- umount /mnt/foo
99
- cephadm.shell:

qa/suites/orch/cephadm/smb/tasks/deploy_smb_basic.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -78,7 +78,7 @@ tasks:
7878
- "client.smbdata"
7979
- cephadm.wait_for_service:
8080
service: smb.saserv1
81-
- cephadm.exec:
81+
- template.exec:
8282
host.b:
8383
- sleep 30
8484
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U smbuser1%insecure321 //{{'host.a'|role_to_remote|attr('ip_address')}}/share1 -c ls"

qa/suites/orch/cephadm/smb/tasks/deploy_smb_ctdb_node_gone_state.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ tasks:
5454
service: smb.modusr1
5555

5656
# verify CTDB is healthy, cluster well formed
57-
- cephadm.exec:
57+
- template.exec:
5858
host.a:
5959
- sleep 30
6060
- "{{ctx.cephadm}} ls --no-detail | {{ctx.cephadm}} shell jq -r 'map(select(.name | startswith(\"smb.modusr1\")))[-1].name' > /tmp/svcname"
@@ -82,7 +82,7 @@ tasks:
8282
service: smb.modusr1
8383

8484
# verify CTDB status doesn't include the node that was removed
85-
- cephadm.exec:
85+
- template.exec:
8686
host.a:
8787
- "{{ctx.cephadm}} ls --no-detail | {{ctx.cephadm}} shell jq -r 'map(select(.name | startswith(\"smb.modusr1\")))[-1].name' > /tmp/svcname"
8888
- "{{ctx.cephadm}} enter -n $(cat /tmp/svcname) ctdb listnodes > /tmp/ctdb_listnodes"

qa/suites/orch/cephadm/smb/tasks/deploy_smb_domain.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -89,7 +89,7 @@ tasks:
8989
- cephadm.wait_for_service:
9090
service: smb.admem1
9191

92-
- cephadm.exec:
92+
- template.exec:
9393
host.b:
9494
- sleep 30
9595
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U DOMAIN1\\\\ckent%1115Rose. //{{'host.a'|role_to_remote|attr('ip_address')}}/share1 -c ls"

qa/suites/orch/cephadm/smb/tasks/deploy_smb_mgr_basic.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -46,7 +46,7 @@ tasks:
4646
- cephadm.wait_for_service:
4747
service: smb.modusr1
4848
# Check if shares exist
49-
- cephadm.exec:
49+
- template.exec:
5050
host.b:
5151
- sleep 30
5252
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user1%t3stP4ss1 //{{'host.a'|role_to_remote|attr('ip_address')}}/share1 -c ls"

qa/suites/orch/cephadm/smb/tasks/deploy_smb_mgr_clustering_ips.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -59,14 +59,14 @@ tasks:
5959
service: smb.modusr1
6060

6161
# Check if shares exist
62-
- cephadm.exec:
62+
- template.exec:
6363
host.d:
6464
- sleep 30
6565
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user1%t3stP4ss1 //{{'host.a'|role_to_remote|attr('ip_address')}}/share1 -c ls"
6666
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user1%t3stP4ss1 //{{'host.a'|role_to_remote|attr('ip_address')}}/share2 -c ls"
6767

6868
# verify CTDB is healthy, cluster well formed
69-
- cephadm.exec:
69+
- template.exec:
7070
host.a:
7171
- "{{ctx.cephadm}} ls --no-detail | {{ctx.cephadm}} shell jq -r 'map(select(.name | startswith(\"smb.modusr1\")))[-1].name' > /tmp/svcname"
7272
- "{{ctx.cephadm}} enter -n $(cat /tmp/svcname) ctdb status > /tmp/ctdb_status"
@@ -78,7 +78,7 @@ tasks:
7878
- rm -rf /tmp/svcname /tmp/ctdb_status
7979

8080
# Test the assigned VIP
81-
- cephadm.exec:
81+
- template.exec:
8282
host.d:
8383
- sleep 30
8484
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user1%t3stP4ss1 //{{VIP0}}/share1 -c ls"

qa/suites/orch/cephadm/smb/tasks/deploy_smb_mgr_ctdb_res_basic.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -88,14 +88,14 @@ tasks:
8888
- cmd: rados --pool=.smb -N uctdb1 get cluster.meta.json /dev/stdout
8989

9090
# Check if shares exist
91-
- cephadm.exec:
91+
- template.exec:
9292
host.d:
9393
- sleep 30
9494
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user1%t3stP4ss1 //{{'host.a'|role_to_remote|attr('ip_address')}}/share1 -c ls"
9595
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user2%t3stP4ss2 //{{'host.a'|role_to_remote|attr('ip_address')}}/share2 -c ls"
9696

9797
# verify CTDB is healthy, cluster well formed
98-
- cephadm.exec:
98+
- template.exec:
9999
host.a:
100100
- "{{ctx.cephadm}} ls --no-detail | {{ctx.cephadm}} shell jq -r 'map(select(.name | startswith(\"smb.uctdb1\")))[-1].name' > /tmp/svcname"
101101
- "{{ctx.cephadm}} enter -n $(cat /tmp/svcname) ctdb status > /tmp/ctdb_status"
@@ -107,7 +107,7 @@ tasks:
107107
- rm -rf /tmp/svcname /tmp/ctdb_status
108108

109109
# Test a different host in the cluster
110-
- cephadm.exec:
110+
- template.exec:
111111
host.d:
112112
- sleep 30
113113
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U user1%t3stP4ss1 //{{'host.c'|role_to_remote|attr('ip_address')}}/share1 -c ls"

qa/suites/orch/cephadm/smb/tasks/deploy_smb_mgr_ctdb_res_dom.yaml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -91,14 +91,14 @@ tasks:
9191
- cmd: rados --pool=.smb -N adctdb1 get cluster.meta.json /dev/stdout
9292

9393
# Check if shares exist
94-
- cephadm.exec:
94+
- template.exec:
9595
host.d:
9696
- sleep 30
9797
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U DOMAIN1\\\\ckent%1115Rose. //{{'host.a'|role_to_remote|attr('ip_address')}}/share1 -c ls"
9898
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U DOMAIN1\\\\ckent%1115Rose. //{{'host.a'|role_to_remote|attr('ip_address')}}/share2 -c ls"
9999

100100
# verify CTDB is healthy, cluster well formed
101-
- cephadm.exec:
101+
- template.exec:
102102
host.a:
103103
- "{{ctx.cephadm}} ls --no-detail | {{ctx.cephadm}} shell jq -r 'map(select(.name | startswith(\"smb.adctdb1\")))[-1].name' > /tmp/svcname"
104104
- "{{ctx.cephadm}} enter -n $(cat /tmp/svcname) ctdb status > /tmp/ctdb_status"
@@ -110,7 +110,7 @@ tasks:
110110
- rm -rf /tmp/svcname /tmp/ctdb_status
111111

112112
# Test a different host in the cluster
113-
- cephadm.exec:
113+
- template.exec:
114114
host.d:
115115
- sleep 30
116116
- "{{ctx.samba_client_container_cmd|join(' ')}} smbclient -U DOMAIN1\\\\ckent%1115Rose. //{{'host.c'|role_to_remote|attr('ip_address')}}/share1 -c ls"

0 commit comments

Comments
 (0)