Skip to content

Commit 9336c9e

Browse files
authored
Merge pull request ceph#62725 from VallariAg/nvmeof-teuthology-fio
qa/suites/nvmeof: Fix thrasher and fio script
2 parents bbed733 + 8217e44 commit 9336c9e

File tree

7 files changed

+57
-6
lines changed

7 files changed

+57
-6
lines changed

qa/suites/nvmeof/basic/clusters/4-gateways-2-initiator.yaml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -29,4 +29,8 @@ overrides:
2929
# cephadm can take up to 5 minutes to bring up remaining mons
3030
mon down mkfs grace: 300
3131
log-ignorelist:
32+
- OSD_DOWN
33+
- OSD_HOST_DOWN
34+
- OSD_ROOT_DOWN
35+
- PG_DEGRADED
3236
- NVMEOF_SINGLE_GATEWAY
Lines changed: 24 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,24 @@
1+
tasks:
2+
- nvmeof:
3+
installer: host.a
4+
gw_image: quay.io/ceph/nvmeof:latest # "default" is the image cephadm defaults to; change to test specific nvmeof images, example "latest"
5+
rbd:
6+
pool_name: mypool
7+
image_name_prefix: myimage
8+
gateway_config:
9+
subsystems_count: 16
10+
namespaces_count: 4 # each subsystem
11+
cli_image: quay.io/ceph/nvmeof-cli:latest
12+
13+
- cephadm.wait_for_service:
14+
service: nvmeof.mypool.mygroup0
15+
16+
- workunit:
17+
no_coverage_and_limits: true
18+
clients:
19+
client.0:
20+
- nvmeof/setup_subsystem.sh
21+
- nvmeof/basic_tests.sh
22+
env:
23+
RBD_POOL: mypool
24+
RBD_IMAGE_PREFIX: myimage

qa/suites/nvmeof/thrash/thrashers/nvmeof_mon_thrash.yaml

Lines changed: 6 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,11 @@
11
overrides:
22
ceph:
3-
log-ignorelist:
3+
log-ignorelist:
4+
# general ignorelist
5+
- OSD_DOWN
6+
- OSD_HOST_DOWN
7+
- OSD_ROOT_DOWN
8+
- PG_DEGRADED
49
# mon thrashing
510
- MON_DOWN
611
- mons down

qa/suites/nvmeof/thrash/thrashers/nvmeof_thrash.yaml

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,11 @@
11
overrides:
22
ceph:
33
log-ignorelist:
4+
# general ignorelist
5+
- OSD_DOWN
6+
- OSD_HOST_DOWN
7+
- OSD_ROOT_DOWN
8+
- PG_DEGRADED
49
# nvmeof daemon thrashing
510
- CEPHADM_FAILED_DAEMON
611
- NVMEOF_SINGLE_GATEWAY

qa/suites/nvmeof/thrash/workloads/fio.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,4 +9,4 @@ tasks:
99
RBD_POOL: mypool
1010
NVMEOF_GROUP: mygroup0
1111
IOSTAT_INTERVAL: '10'
12-
RUNTIME: '1800'
12+
RUNTIME: '1200'

qa/tasks/nvmeof.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -321,7 +321,7 @@ def __init__(self, ctx, config, daemons) -> None:
321321
self.max_thrash_daemons = int(self.config.get('max_thrash', len(self.daemons) - 1))
322322

323323
# Limits on thrashing each daemon
324-
self.daemon_max_thrash_times = int(self.config.get('daemon_max_thrash_times', 5))
324+
self.daemon_max_thrash_times = int(self.config.get('daemon_max_thrash_times', 4))
325325
self.daemon_max_thrash_period = int(self.config.get('daemon_max_thrash_period', 30 * 60)) # seconds
326326

327327
self.min_thrash_delay = int(self.config.get('min_thrash_delay', 60))
@@ -436,7 +436,7 @@ def revive_daemon(self, daemon, killed_method):
436436
if killed_method == "ceph_daemon_stop":
437437
daemon.remote.run(args=[
438438
"ceph", "orch", "daemon", "restart",
439-
name
439+
name, "--force"
440440
])
441441
# note: temporarily use 'daemon start' to restart
442442
# daemons instead of 'systemctl start'

qa/workunits/nvmeof/fio_test.sh

Lines changed: 15 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -54,20 +54,31 @@ filename=$(echo "$selected_drives" | sed -z 's/\n/:\/dev\//g' | sed 's/:\/dev\/$
5454
filename="/dev/$filename"
5555

5656
cat >> $fio_file <<EOF
57-
[nvmeof-fio-test]
57+
[global]
5858
ioengine=${IO_ENGINE:-sync}
5959
bsrange=${BS_RANGE:-4k-64k}
6060
numjobs=${NUM_OF_JOBS:-1}
6161
size=${SIZE:-1G}
6262
time_based=1
6363
runtime=$RUNTIME
6464
rw=${RW:-randrw}
65-
filename=${filename}
6665
verify=md5
6766
verify_fatal=1
67+
do_verify=1
68+
serialize_overlap=1
69+
group_reporting
6870
direct=1
71+
6972
EOF
7073

74+
for i in $selected_drives; do
75+
echo "[job-$i]" >> "$fio_file"
76+
echo "filename=/dev/$i" >> "$fio_file"
77+
echo "" >> "$fio_file" # Adds a blank line
78+
done
79+
80+
cat $fio_file
81+
7182
status_log() {
7283
POOL="${RBD_POOL:-mypool}"
7384
GROUP="${NVMEOF_GROUP:-mygroup0}"
@@ -79,6 +90,7 @@ status_log() {
7990
ceph nvme-gw show $POOL $GROUP
8091
sudo nvme list
8192
sudo nvme list | wc -l
93+
sudo nvme list-subsys
8294
for device in $selected_drives; do
8395
echo "Processing device: $device"
8496
sudo nvme list-subsys /dev/$device
@@ -103,6 +115,7 @@ fio --showcmd $fio_file
103115
set +e
104116
sudo fio $fio_file
105117
if [ $? -ne 0 ]; then
118+
echo "[nvmeof.fio]: fio failed!"
106119
status_log
107120
exit 1
108121
fi

0 commit comments

Comments
 (0)