Search is not available for this dataset
repo
stringlengths
2
152
file
stringlengths
15
239
code
stringlengths
0
58.4M
file_length
int64
0
58.4M
avg_line_length
float64
0
1.81M
max_line_length
int64
0
12.7M
extension_type
stringclasses
364 values
null
ceph-main/qa/suites/krbd/wac/wac/clusters/fixed-3.yaml
.qa/clusters/fixed-3.yaml
25
25
25
yaml
null
ceph-main/qa/suites/krbd/wac/wac/tasks/wac.yaml
tasks: - exec: client.0: - "dmesg -C" - workunit: clients: all: - rbd/krbd_wac.sh
108
11.111111
25
yaml
null
ceph-main/qa/suites/krbd/wac/wac/verify/many-resets.yaml
overrides: ceph: conf: global: ms inject socket failures: 500 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME tasks: - exec: client.0: - "dmesg | grep -q 'libceph: osd.* socket closed'" - "dmesg | grep -q 'libceph: osd.* socket error on write...
323
22.142857
62
yaml
null
ceph-main/qa/suites/krbd/wac/wac/verify/no-resets.yaml
tasks: - exec: client.0: - "! dmesg | grep -q 'libceph: osd.* socket closed'" - "! dmesg | grep -q 'libceph: osd.* socket error on write'"
151
24.333333
64
yaml
null
ceph-main/qa/suites/mixed-clients/basic/clusters/fixed-3.yaml
roles: - [mon.a, mgr.x, mds.a, osd.0, osd.1] - [mon.b, mon.c, osd.2, osd.3, client.0] - [client.1]
99
19
40
yaml
null
ceph-main/qa/suites/mixed-clients/basic/tasks/kernel_cfuse_workunits_dbench_iozone.yaml
overrides: ceph: conf: global: ms die on skipped message: false tasks: - install: branch: dumpling - ceph: - parallel: - user-workload - kclient-workload user-workload: sequential: - ceph-fuse: [client.0] - workunit: clients: client.0: - suites/iozone.sh kcl...
452
15.777778
40
yaml
null
ceph-main/qa/suites/mixed-clients/basic/tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml
overrides: ceph: conf: global: ms die on skipped message: false tasks: - install: branch: dumpling - ceph: - parallel: - user-workload - kclient-workload user-workload: sequential: - ceph-fuse: [client.0] - workunit: clients: client.0: - suites/blogbench.sh ...
459
16.037037
40
yaml
null
ceph-main/qa/suites/netsplit/ceph.yaml
overrides: ceph: conf: global: mon election default strategy: 3 mon: mon min osdmap epochs: 25 paxos service trim min: 5 # prune full osdmaps regularly mon osdmap full prune min: 15 mon osdmap full prune interval: 2 mon osdmap full prune txsize: ...
593
23.75
63
yaml
null
ceph-main/qa/suites/netsplit/cluster.yaml
roles: - [mon.a, mgr.x, osd.0, osd.1, osd.2, osd.3] - [mon.b, mgr.y, osd.4, osd.5, osd.6, osd.7, client.0] - [mon.c] openstack: - volumes: # attached to each instance count: 4 size: 10 # GB overrides: ceph: conf: osd: osd shutdown pgref assert: true
278
18.928571
54
yaml
null
ceph-main/qa/suites/netsplit/msgr.yaml
../../msgr/async.yaml
21
21
21
yaml
null
ceph-main/qa/suites/netsplit/rados.yaml
.qa/config/rados.yaml
21
21
21
yaml
null
ceph-main/qa/suites/netsplit/tests/mon_pool_ops.yaml
overrides: ceph: conf: global: mon election default strategy: 3 tasks: - workunit: clients: client.0: - mon/pool_ops.sh - netsplit.disconnect: [mon.a, mon.c] - workunit: clients: client.0: - mon/pool_ops.sh - netsplit.reconnect: [mon.a, mon.c] - netsplit.disconnec...
404
18.285714
40
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/0-centos_8.stream_container_tools.yaml
.qa/distros/podman/centos_8.stream_container_tools.yaml
55
55
55
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/1-start.yaml
tasks: - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls roles: - - host.a - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - - host.b - osd.4 - osd.5 - osd.6 - osd.7 openstack: - volumes: # attached to each ...
440
13.7
39
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/2-nfs.yaml
tasks: # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs - cephadm.wait_for_service: service: mds.foofs - cephadm.shell: host.a: - ceph nfs cluster create foo --placement=2 || ceph nfs clus...
879
28.333333
170
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/3-upgrade-with-workload.yaml
tasks: - parallel: - upgrade-tasks - workload-tasks upgrade-tasks: sequential: - cephadm.shell: env: [sha1] host.a: - ceph config set mon mon_warn_on_insecure_global_id_reclaim false --force - ceph config set mon mon_warn_on_insecure_global_id_reclaim_allowed false --force -...
1,576
34.840909
231
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/4-final.yaml
tasks: - vip.exec: host.a: - umount /mnt/foo - cephadm.shell: host.a: - ceph nfs cluster ls | grep foo - ceph nfs export ls foo --detailed - rados -p .nfs --all ls - - ceph config get mgr mgr/cephadm/migration_current | grep 6
265
23.181818
66
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/1-bootstrap/16.2.0.yaml
tasks: - cephadm: roleless: true image: quay.io/ceph/ceph:v16.2.0 cephadm_branch: v16.2.0 cephadm_git_url: https://github.com/ceph/ceph # needed for v16.2.0 due to --skip-admin-label avoid_pacific_features: true
236
25.333333
50
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/1-bootstrap/16.2.4.yaml
tasks: - cephadm: roleless: true image: quay.io/ceph/ceph:v16.2.4 cephadm_branch: v16.2.4 cephadm_git_url: https://github.com/ceph/ceph # needed for v16.2.4 due to --skip-admin-label avoid_pacific_features: true
236
25.333333
50
yaml
null
ceph-main/qa/suites/orch/cephadm/mgr-nfs-upgrade/1-bootstrap/16.2.5.yaml
tasks: - cephadm: roleless: true image: quay.io/ceph/ceph:v16.2.5 cephadm_branch: v16.2.5 cephadm_git_url: https://github.com/ceph/ceph
152
20.857143
49
yaml
null
ceph-main/qa/suites/orch/cephadm/orchestrator_cli/2-node-mgr.yaml
.qa/clusters/2-node-mgr.yaml
28
28
28
yaml
null
ceph-main/qa/suites/orch/cephadm/orchestrator_cli/orchestrator_cli.yaml
tasks: - install: - ceph: # tests may leave mgrs broken, so don't try and call into them # to invoke e.g. pg dump during teardown. wait-for-scrub: false log-ignorelist: - overall HEALTH_ - \(MGR_DOWN\) - \(DEVICE_IDENT_ON\) - \(DEVICE_FAULT_ON\) - \(P...
478
25.611111
68
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/0-nvme-loop.yaml
.qa/overrides/nvme_loop.yaml
28
28
28
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/1-start.yaml
tasks: - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices ' roles: - - host.a - client.0 - - host.b - client.1 openstack: - volumes...
461
16.769231
57
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/2-ops/repave-all.yaml
tasks: - cephadm.shell: host.a: - | set -e set -x ceph orch ps ceph orch device ls ceph osd tree for osd in `ceph osd ls` ; do ceph orch osd rm $osd --force --zap --replace done while ceph orch osd rm ls | wc | grep ^1 ; do sleep 10 ;...
326
22.357143
69
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/2-ops/rm-zap-add.yaml
tasks: - cephadm.shell: host.a: - | set -e set -x ceph orch ps ceph orch device ls DEVID=$(ceph device ls | grep osd.1 | awk '{print $1}') HOST=$(ceph orch device ls | grep $DEVID | awk '{print $1}') DEV=$(ceph orch device ls | grep $DEVID | awk '{print ...
636
34.388889
72
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/2-ops/rm-zap-flag.yaml
tasks: - cephadm.shell: host.a: - | set -e set -x ceph orch ps ceph orch device ls DEVID=$(ceph device ls | grep osd.1 | awk '{print $1}') HOST=$(ceph orch device ls | grep "$DEVID" | awk '{print $1}') DEV=$(ceph orch device ls | grep "$DEVID" | awk '{pr...
571
34.75
79
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/2-ops/rm-zap-wait.yaml
tasks: - cephadm.shell: host.a: - | set -e set -x ceph orch ps ceph orch device ls DEVID=$(ceph device ls | grep osd.1 | awk '{print $1}') HOST=$(ceph orch device ls | grep $DEVID | awk '{print $1}') DEV=$(ceph orch device ls | grep $DEVID | awk '{print ...
592
33.882353
72
yaml
null
ceph-main/qa/suites/orch/cephadm/osds/2-ops/rmdir-reactivate.yaml
tasks: - cephadm.shell: host.a: - | set -e set -x ceph orch ps HOST=$(hostname -s) OSD=$(ceph orch ps $HOST | grep osd | head -n 1 | awk '{print $1}') echo "host $HOST, osd $OSD" ceph orch daemon stop $OSD while ceph orch ps | grep $OSD | grep ru...
678
31.333333
75
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/0-nvme-loop.yaml
.qa/overrides/nvme_loop.yaml
28
28
28
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/1-start.yaml
tasks: - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls roles: - - host.a - client.0 - - host.b - client.1 openstack: - volumes: # attached to each instance count: 4 size: 10 # ...
403
15.16
39
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/3-final.yaml
tasks: - cephadm.shell: host.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices '
270
23.636364
57
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/basic.yaml
0
0
0
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/client-keyring.yaml
tasks: - cephadm.shell: host.a: - ceph orch host label add `hostname` foo - ceph auth get-or-create client.foo mon 'allow r' - ceph orch client-keyring set client.foo label:foo --mode 770 --owner 11111:22222 - exec: host.a: - while ! test -e /etc/ceph/ceph.client.foo.keyring ; do sleep 1...
1,405
33.292683
88
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/iscsi.yaml
tasks: - cephadm.shell: host.a: - ceph osd pool create foo - rbd pool init foo - ceph orch apply iscsi foo u p - cephadm.wait_for_service: service: iscsi.foo
184
19.555556
37
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/jaeger.yaml
tasks: - cephadm.shell: host.a: - ceph orch apply jaeger - cephadm.wait_for_service: service: elasticsearch - cephadm.wait_for_service: service: jaeger-collector - cephadm.wait_for_service: service: jaeger-query - cephadm.wait_for_service: service: jaeger-agent
287
23
30
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/mirror.yaml
tasks: - cephadm.shell: host.a: - ceph orch apply rbd-mirror "--placement=*" - ceph orch apply cephfs-mirror "--placement=*" - cephadm.wait_for_service: service: rbd-mirror - cephadm.wait_for_service: service: cephfs-mirror
248
23.9
53
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs-haproxy-proto.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server # use nfs module to create cluster and export - cephadm.shell: host.a: - ceph fs volume c...
955
25.555556
101
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs-ingress-rgw-bucket.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph orch apply rgw foorgw --port 8800 - ceph nfs cluste...
2,317
24.755556
119
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs-ingress-rgw-user.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph orch apply rgw foorgw --port 8800 - ceph nfs cluste...
2,364
24.989011
119
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs-ingress.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs # deploy nfs + ingress - cephadm....
1,716
23.884058
89
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs-ingress2.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs - ceph nfs cluster create fo...
2,167
29.535211
96
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs-keepalive-only.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph fs volume create foofs # deploy nfs + keepalive-only ing...
1,308
22.375
89
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs.yaml
tasks: # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.apply: specs: - service_type: nfs service_id: foo - cephadm.wait_for_service: service: nfs.foo
234
15.785714
36
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/nfs2.yaml
tasks: # stop kernel nfs server, if running - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph nfs cluster create foo - cephadm.wait_for_service: service: nfs.foo
221
16.076923
36
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/rgw-ingress.yaml
tasks: - vip: # make sure cephadm notices the new IP - cephadm.shell: host.a: - ceph orch device ls --refresh # deploy rgw + ingress - cephadm.apply: specs: - service_type: rgw service_id: foo placement: count: 4 host_pattern: "*" spec: rgw_fro...
1,907
30.278689
85
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-roleless/2-services/rgw.yaml
tasks: - cephadm.apply: specs: - service_type: rgw service_id: foo placement: count_per_host: 4 host_pattern: "*" spec: rgw_frontend_port: 8000 - cephadm.wait_for_service: service: rgw.foo
257
18.846154
33
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-singlehost/1-start.yaml
tasks: - cephadm: roleless: true single_host_defaults: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls roles: - - host.a - osd.0 - osd.1 - osd.2 - osd.3 - client.0 openstack: - volumes: # attache...
450
15.107143
39
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-singlehost/3-final.yaml
tasks: - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls
157
16.555556
27
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-singlehost/2-services/basic.yaml
0
0
0
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-singlehost/2-services/rgw.yaml
tasks: - cephadm.apply: specs: - service_type: rgw service_id: foo placement: count_per_host: 4 host_pattern: "*" spec: rgw_frontend_port: 8000 - cephadm.wait_for_service: service: rgw.foo
257
18.846154
33
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-small/0-nvme-loop.yaml
.qa/overrides/nvme_loop.yaml
28
28
28
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-small/fixed-2.yaml
roles: - - mon.a - mgr.y - osd.0 - client.0 - ceph.rgw.foo.a - node-exporter.a - alertmanager.a - - mon.b - mgr.x - osd.1 - client.1 - prometheus.a - grafana.a - node-exporter.b - - mon.c - mgr.z - osd.2 - client.2 - node-exporter.c openstack: - volumes: # attached to each instance c...
426
13.233333
39
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-small/start.yaml
tasks: - cephadm: conf: mgr: debug ms: 1 debug mgr: 20 - cephadm.shell: mon.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls --format yaml ...
356
20
54
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-small/0-distro/centos_8.stream_container_tools_crun.yaml
../.qa/distros/container-hosts/centos_8.stream_container_tools_crun.yaml
72
72
72
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-small/agent/off.yaml
overrides: ceph: conf: mgr: mgr/cephadm/use_agent: false
77
12
36
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke-small/agent/on.yaml
overrides: ceph: conf: mgr: mgr/cephadm/use_agent: true
76
11.833333
35
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke/0-nvme-loop.yaml
.qa/overrides/nvme_loop.yaml
28
28
28
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke/fixed-2.yaml
roles: - - mon.a - mon.c - mgr.y - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - ceph.rgw.foo.a - node-exporter.a - alertmanager.a - - mon.b - mgr.x - osd.4 - osd.5 - osd.6 - osd.7 - client.1 - prometheus.a - grafana.a - node-exporter.b - ceph.iscsi.iscsi.a openstack: - volumes: # a...
456
12.848485
39
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke/start.yaml
tasks: - cephadm: conf: mgr: debug ms: 1 debug mgr: 20 - cephadm.shell: mon.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls --format yaml ...
356
20
54
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke/agent/off.yaml
overrides: ceph: conf: mgr: mgr/cephadm/use_agent: false
77
12
36
yaml
null
ceph-main/qa/suites/orch/cephadm/smoke/agent/on.yaml
overrides: ceph: conf: mgr: mgr/cephadm/use_agent: true
76
11.833333
35
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/1-start.yaml
tasks: - install: - cephadm: conf: mgr: debug ms: 1 debug mgr: 20
92
10.625
21
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/2-thrash.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost conf: osd: osd debug reject backfill probability: .3 osd scrub min interval: 60 osd scrub max interval: 120 osd max backfills: 3 osd snap trim sleep: 2 ...
696
24.814815
49
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/fixed-2.yaml
../smoke/fixed-2.yaml
21
21
21
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/root.yaml
overrides: cephadm: cephadm_mode: root
45
10.5
22
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/3-tasks/rados_api_tests.yaml
.qa/suites/rados/thrash/workloads/rados_api_tests.yaml
54
54
54
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/3-tasks/radosbench.yaml
.qa/suites/rados/thrash/workloads/radosbench.yaml
49
49
49
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/3-tasks/small-objects.yaml
.qa/suites/rados/thrash/workloads/small-objects.yaml
52
52
52
yaml
null
ceph-main/qa/suites/orch/cephadm/thrash/3-tasks/snaps-few-objects.yaml
.qa/suites/rados/thrash/workloads/snaps-few-objects.yaml
56
56
56
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/4-wait.yaml
tasks: - cephadm.shell: env: [sha1] mon.a: - while ceph orch upgrade status | jq '.in_progress' | grep true && ! ceph orch upgrade status | jq '.message' | grep Error ; do ceph orch ps ; ceph versions ; ceph orch upgrade status ; ceph health detail ; sleep 30 ; done - ceph orch ps - ceph versi...
666
38.235294
229
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/5-upgrade-ls.yaml
tasks: - cephadm.shell: mon.a: - ceph orch upgrade ls - ceph orch upgrade ls --image quay.io/ceph/ceph --show-all-versions | grep 16.2.0 - ceph orch upgrade ls --image quay.io/ceph/ceph --tags | grep v16.2.2
230
32
88
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/1-start-distro/1-start-centos_8.stream_container-tools.yaml
os_type: centos os_version: "8.stream" tasks: - pexec: all: - sudo cp /etc/containers/registries.conf /etc/containers/registries.conf.backup - sudo dnf -y module reset container-tools - sudo dnf -y module install container-tools - sudo cp /etc/containers/registries.conf.backup /etc/containers/reg...
841
20.05
84
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/1-start-distro/1-start-ubuntu_20.04.yaml
os_type: ubuntu os_version: "20.04" tasks: - cephadm: image: quay.io/ceph/ceph:v16.2.0 cephadm_branch: v16.2.0 cephadm_git_url: https://github.com/ceph/ceph # avoid --cap-add=PTRACE + --privileged for older cephadm versions allow_ptrace: false avoid_pacific_features: true roles: - - mon.a - ...
552
15.264706
70
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/2-repo_digest/defaut.yaml
0
0
0
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/2-repo_digest/repo_digest.yaml
tasks: - cephadm.shell: mon.a: - ceph config set mgr mgr/cephadm/use_repo_digest false --force
105
20.2
69
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/3-upgrade/simple.yaml
tasks: - cephadm.shell: env: [sha1] mon.a: # setup rgw - radosgw-admin realm create --rgw-realm=r --default - radosgw-admin zonegroup create --rgw-zonegroup=default --master --default - radosgw-admin zone create --rgw-zonegroup=default --rgw-zone=z --master --default - radosgw-admi...
1,007
44.818182
99
yaml
null
ceph-main/qa/suites/orch/cephadm/upgrade/3-upgrade/staggered.yaml
tasks: - cephadm.shell: env: [sha1] mon.a: # setup rgw - radosgw-admin realm create --rgw-realm=r --default - radosgw-admin zonegroup create --rgw-zonegroup=default --master --default - radosgw-admin zone create --rgw-zonegroup=default --rgw-zone=z --master --default - radosgw-admi...
7,926
58.601504
208
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/fixed-2.yaml
../smoke/fixed-2.yaml
21
21
21
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/start.yaml
tasks: - install: - cephadm: conf: mgr: debug ms: 1 debug mgr: 20
92
10.625
21
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/mode/packaged.yaml
overrides: cephadm: cephadm_mode: cephadm-package install: extra_packages: [cephadm]
97
15.333333
33
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/mode/root.yaml
overrides: cephadm: cephadm_mode: root
45
10.5
22
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/tasks/rados_api_tests.yaml
.qa/suites/rados/basic/tasks/rados_api_tests.yaml
49
49
49
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/tasks/rados_python.yaml
.qa/suites/rados/basic/tasks/rados_python.yaml
46
46
46
yaml
null
ceph-main/qa/suites/orch/cephadm/with-work/tasks/rotate-keys.yaml
tasks: - cephadm.shell: mon.a: - | set -ex for f in osd.0 osd.1 osd.2 osd.3 osd.4 osd.5 osd.6 osd.7 mgr.y mgr.x do echo "rotating key for $f" K=$(ceph auth get-key $f) NK="$K" ceph orch daemon rotate-key $f while [ "$K" == "$NK" ]; do ...
400
22.588235
74
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_adoption.yaml
roles: - [mon.a, mgr.x, osd.0, client.0] tasks: - install: - exec: mon.a: - yum install -y python3 || apt install -y python3 - workunit: clients: client.0: - cephadm/test_adoption.sh
211
16.666667
56
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_cephadm.yaml
roles: - [mon.a, mgr.x, osd.0, client.0] tasks: - install: - exec: mon.a: - yum install -y python3 || apt install -y python3 - workunit: clients: client.0: - cephadm/test_cephadm.sh
210
16.583333
56
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_cephadm_repos.yaml
roles: - [mon.a, mgr.x, osd.0, client.0] tasks: - workunit: no_coverage_and_limits: true clients: client.0: - cephadm/test_repos.sh
154
16.222222
33
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_extra_daemon_features.yaml
roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 tasks: - install: - cephadm: - exec: all-hosts: - mkdir /etc/cephadm_testing - cephadm.apply: specs: - service_type: mon placement: host_pattern: '*' extra_container_args: - "...
2,257
29.106667
86
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_nfs.yaml
roles: - - host.a - osd.0 - osd.1 - osd.2 - mon.a - mgr.a - client.0 tasks: - install: - cephadm: - cephadm.shell: host.a: - ceph orch apply mds a - cephfs_test_runner: modules: - tasks.cephfs.test_nfs
234
12.055556
29
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_orch_cli.yaml
roles: - - host.a - osd.0 - osd.1 - osd.2 - mon.a - mgr.a - client.0 tasks: - install: - cephadm: - cephadm.shell: host.a: - ceph orch apply mds a - cephfs_test_runner: modules: - tasks.cephadm_cases.test_cli
241
12.444444
36
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_orch_cli_mon.yaml
roles: - - host.a - osd.0 - osd.1 - osd.2 - mon.a - mgr.a - client.0 - - host.b - osd.3 - osd.4 - osd.5 - mon.b - mgr.b - client.1 - - host.c - osd.6 - osd.7 - osd.8 - mon.c - mgr.c - client.2 - - host.d - osd.9 - osd.10 - osd.11 - mon.d - mgr.d - client.3 - - host.e - ...
546
10.891304
40
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_rgw_multisite.yaml
roles: - - host.a - mon.a - mgr.a - osd.0 - - host.b - mon.b - mgr.b - osd.1 - - host.c - mon.c - osd.2 tasks: - install: - cephadm: - cephadm.shell: host.a: - ceph mgr module enable rgw - rgw_module.apply: specs: - rgw_realm: myrealm1 rgw_zonegroup: myzonegroup1 rgw_...
1,257
29.682927
168
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_set_mon_crush_locations.yaml
roles: - - host.a - osd.0 - mon.a - mgr.a - - host.b - osd.1 - mon.b - mgr.b - - host.c - osd.2 - mon.c tasks: - install: - cephadm: - cephadm.apply: specs: - service_type: mon service_id: foo placement: count: 3 spec: crush_locations: ho...
1,999
30.746032
164
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_iscsi_container/centos_8.stream_container_tools.yaml
.qa/distros/podman/centos_8.stream_container_tools.yaml
55
55
55
yaml
null
ceph-main/qa/suites/orch/cephadm/workunits/task/test_iscsi_container/test_iscsi_container.yaml
roles: - - host.a - osd.0 - osd.1 - osd.2 - mon.a - mgr.a - client.0 tasks: - install: - cephadm: - cephadm.shell: host.a: - ceph osd pool create foo - rbd pool init foo - ceph orch apply iscsi foo u p - workunit: clients: client.0: - cephadm/test_iscsi_pids_limit.sh ...
362
15.5
42
yaml
null
ceph-main/qa/suites/orch/rook/smoke/0-kubeadm.yaml
tasks: - kubeadm:
18
5.333333
10
yaml
null
ceph-main/qa/suites/orch/rook/smoke/0-nvme-loop.yaml
.qa/overrides/nvme_loop.yaml
28
28
28
yaml
null
ceph-main/qa/suites/orch/rook/smoke/1-rook.yaml
tasks: - rook:
15
4.333333
7
yaml
null
ceph-main/qa/suites/orch/rook/smoke/0-distro/ubuntu_20.04.yaml
.qa/distros/container-hosts/ubuntu_20.04.yaml
45
45
45
yaml