Search is not available for this dataset
repo stringlengths 2 152 ⌀ | file stringlengths 15 239 | code stringlengths 0 58.4M | file_length int64 0 58.4M | avg_line_length float64 0 1.81M | max_line_length int64 0 12.7M | extension_type stringclasses 364
values |
|---|---|---|---|---|---|---|
null | ceph-main/qa/suites/teuthology/no-ceph/tasks/teuthology.yaml | tasks:
- tests:
| 20 | 6 | 12 | yaml |
null | ceph-main/qa/suites/teuthology/nop/clusters/empty.yaml | 0 | 0 | 0 | yaml | |
null | ceph-main/qa/suites/teuthology/nop/clusters/single.yaml | roles:
- [mon.a, mgr.x, client.0]
| 38 | 12 | 30 | yaml |
null | ceph-main/qa/suites/teuthology/nop/tasks/nop.yaml | tasks:
- nop:
| 19 | 4 | 10 | yaml |
null | ceph-main/qa/suites/teuthology/rgw/tasks/s3tests-fastcgi.yaml | # this runs s3tests against rgw, using mod_fastcgi
roles:
- [mon.a, mon.c, osd.0, osd.1, osd.2, client.0]
- [mon.b, mgr.x, osd.3, osd.4, osd.5, client.1]
tasks:
- install:
branch: master
- ceph:
- rgw: [client.0]
- s3tests:
client.0:
rgw_server: client.0
force-branch: ceph-master
overrides:
ceph:... | 465 | 17.64 | 50 | yaml |
null | ceph-main/qa/suites/teuthology/rgw/tasks/s3tests-fcgi.yaml | # this runs s3tests against rgw, using mod_proxy_fcgi
# the choice between uds or tcp with mod_proxy_fcgi depends on the distro
roles:
- [mon.a, mon.c, osd.0, osd.1, osd.2, client.0]
- [mon.b, mgr.x, osd.3, osd.4, osd.5, client.1]
tasks:
- install:
branch: master
- ceph:
- rgw: [client.0]
- s3tests:
client.0:
... | 561 | 19.814815 | 73 | yaml |
null | ceph-main/qa/suites/teuthology/rgw/tasks/s3tests.yaml | # this runs s3tests against rgw
roles:
- [mon.a, mon.c, osd.0, osd.1, osd.2, client.0]
- [mon.b, mgr.x, osd.3, osd.4, osd.5, client.1]
tasks:
- install:
branch: master
- ceph:
- rgw: [client.0]
- s3tests:
client.0:
rgw_server: client.0
force-branch: ceph-master
overrides:
ceph:
fs: xfs
co... | 425 | 16.75 | 47 | yaml |
null | ceph-main/qa/suites/teuthology/workunits/yes.yaml | roles:
- [client.0]
tasks:
- install:
- workunit:
clients:
all:
- true.sh
| 96 | 9.777778 | 17 | yaml |
null | ceph-main/qa/suites/tgt/basic/clusters/fixed-3.yaml | roles:
- [mon.a, mon.c, osd.0, osd.1, osd.2]
- [mon.b, mgr.x, mds.a, osd.3, osd.4, osd.5]
- [client.0]
| 103 | 19.8 | 44 | yaml |
null | ceph-main/qa/suites/tgt/basic/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/tgt/basic/msgr-failures/many.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 500
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 176 | 18.666667 | 44 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/blogbench.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/blogbench.sh
| 114 | 10.5 | 33 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/bonnie.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/bonnie.sh
| 111 | 10.2 | 30 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/dbench-short.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/dbench-short.sh
| 117 | 10.8 | 36 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/dbench.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/dbench.sh
| 111 | 10.2 | 30 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/ffsb.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/ffsb.sh
| 109 | 10 | 28 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/fio.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/fio.sh
| 108 | 9.9 | 27 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/fsstress.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/fsstress.sh
| 113 | 10.4 | 32 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/fsx.yaml | tasks:
- install:
extra_system_packages:
deb:
- libaio-dev
- libtool-bin
- uuid-dev
- xfslibs-dev
rpm:
- libaio-devel
- libtool
- libuuid-devel
- xfsprogs-devel
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/fsx.sh
| 315 | 14.047619 | 27 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/fsync-tester.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/fsync-tester.sh
| 117 | 10.8 | 36 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/iogen.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/iogen.sh
| 110 | 10.1 | 29 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/iozone-sync.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/iozone-sync.sh
| 116 | 10.7 | 35 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/iozone.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/iozone.sh
| 111 | 10.2 | 30 | yaml |
null | ceph-main/qa/suites/tgt/basic/tasks/pjd.yaml | tasks:
- install:
- ceph:
- tgt:
- iscsi:
- workunit:
clients:
all:
- suites/pjd.sh
| 108 | 9.9 | 27 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/0-start.yaml | roles:
- - mon.a
- mon.c
- mgr.y
- mds.a
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- node-exporter.a
- alertmanager.a
- - mon.b
- mds.b
- mgr.x
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
- prometheus.a
- grafana.a
- node-exporter.b
openstack:
- volumes: # attached to each instan... | 460 | 12.558824 | 39 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/1-tasks.yaml | tasks:
- install:
branch: pacific
exclude_packages:
- ceph-volume
- print: "**** done install task..."
- print: "**** done start installing pacific cephadm ..."
- cephadm:
image: quay.ceph.io/ceph-ci/ceph:pacific
cephadm_branch: pacific
cephadm_git_url: https://github.com/ceph/ceph
conf:
... | 1,211 | 26.545455 | 80 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/upgrade-sequence.yaml | # renamed tasks: to upgrade-sequence:
upgrade-sequence:
sequential:
- print: "**** done start upgrade, wait"
- cephadm.shell:
env: [sha1]
mon.a:
- ceph config set global log_to_journald false --force
- ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1
- wh... | 650 | 37.294118 | 125 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/workload/ec-rados-default.yaml | meta:
- desc: |
run run randomized correctness test for rados operations
on an erasure-coded pool
workload:
full_sequential:
- print: "**** done start ec-rados-default.yaml"
- rados:
clients: [client.0]
ops: 4000
objects: 50
ec_pool: true
write_append_excl: false
op_weigh... | 577 | 21.230769 | 59 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/workload/rados_api.yaml | meta:
- desc: |
object class functional tests
workload:
full_sequential:
- print: "**** done start rados_api.yaml"
- workunit:
branch: pacific
clients:
client.0:
- cls
- print: "**** done end rados_api.yaml"
| 263 | 19.307692 | 45 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/workload/rados_loadgenbig.yaml | meta:
- desc: |
generate read/write load with rados objects ranging from 1MB to 25MB
workload:
full_sequential:
- print: "**** done start rados_loadgenbig.yaml"
- workunit:
branch: pacific
clients:
client.0:
- rados/load-gen-big.sh
- print: "**** done end rados_loa... | 334 | 24.769231 | 71 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/workload/rbd_import_export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
workload:
full_sequential:
- print: "**** done start rbd_import_export.yaml"
- workunit:
branch: pacific
clients:
client.1:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- pri... | 363 | 23.266667 | 53 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/workload/test_rbd_api.yaml | meta:
- desc: |
librbd C and C++ api tests
workload:
full_sequential:
- print: "**** done start test_rbd_api.yaml"
- workunit:
branch: pacific
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done end test_rbd_api.yaml"
| 283 | 20.846154 | 48 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/parallel/workload/test_rbd_python.yaml | meta:
- desc: |
librbd python api tests
workload:
full_sequential:
- print: "**** done start test_rbd_python.yaml"
- workunit:
branch: pacific
clients:
client.0:
- rbd/test_librbd_python.sh
- print: "**** done end test_rbd_python.yaml"
| 292 | 19.928571 | 51 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/0-roles.yaml | roles:
- - mon.a
- mon.c
- mgr.y
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- node-exporter.a
- alertmanager.a
- - mon.b
- mgr.x
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
- prometheus.a
- grafana.a
- node-exporter.b
openstack:
- volumes: # attached to each instance
count: 4
... | 440 | 12.78125 | 39 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/1-start.yaml | tasks:
- install:
branch: pacific
exclude_packages:
- ceph-volume
- cephadm:
image: quay.ceph.io/ceph-ci/ceph:pacific
cephadm_branch: pacific
cephadm_git_url: https://github.com/ceph/ceph
conf:
osd:
#set config option for which cls modules are allowed to be loaded / used
... | 3,188 | 24.926829 | 126 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/2-first-half-tasks/radosbench.yaml | meta:
- desc: |
run randomized correctness test for rados operations
generate write load with rados bench
first-half-tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
... | 418 | 19.95 | 55 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/2-first-half-tasks/rbd-cls.yaml | meta:
- desc: |
run basic cls tests for rbd
first-half-tasks:
- workunit:
branch: pacific
clients:
client.0:
- cls/test_cls_rbd.sh
- print: "**** done cls/test_cls_rbd.sh 5-workload"
| 208 | 18 | 51 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/2-first-half-tasks/rbd-import-export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
first-half-tasks:
- workunit:
branch: pacific
clients:
client.0:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 5-workload"
| 269 | 19.769231 | 52 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/2-first-half-tasks/rbd_api.yaml | meta:
- desc: |
librbd C and C++ api tests
first-half-tasks:
- workunit:
branch: pacific
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done rbd/test_librbd.sh 7-workload"
| 212 | 18.363636 | 50 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/2-first-half-tasks/readwrite.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool,
using only reads, writes, and deletes
first-half-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 500
write_append_excl: false
op_weights:
read: 45
write:... | 391 | 22.058824 | 73 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/2-first-half-tasks/snaps-few-objects.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool with snapshot operations
first-half-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 50
write_append_excl: false
op_weights:
read: 100
write: 100
dele... | 452 | 22.842105 | 97 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/3-stress-tasks/radosbench.yaml | meta:
- desc: |
run randomized correctness test for rados operations
generate write load with rados bench
stress-tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- ... | 528 | 19.346154 | 55 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/3-stress-tasks/rbd-cls.yaml | meta:
- desc: |
run basic cls tests for rbd
stress-tasks:
- workunit:
branch: pacific
clients:
client.0:
- cls/test_cls_rbd.sh
- print: "**** done cls/test_cls_rbd.sh 5-workload"
| 204 | 17.636364 | 51 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/3-stress-tasks/rbd-import-export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
stress-tasks:
- workunit:
branch: pacific
clients:
client.0:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 5-workload"
| 265 | 19.461538 | 52 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/3-stress-tasks/rbd_api.yaml | meta:
- desc: |
librbd C and C++ api tests
stress-tasks:
- workunit:
branch: pacific
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done rbd/test_librbd.sh 7-workload"
| 208 | 18 | 50 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/3-stress-tasks/readwrite.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool,
using only reads, writes, and deletes
stress-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 500
write_append_excl: false
op_weights:
read: 45
write: 45
... | 387 | 21.823529 | 73 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/3-stress-tasks/snaps-few-objects.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool with snapshot operations
stress-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 50
write_append_excl: false
op_weights:
read: 100
write: 100
delete: ... | 448 | 22.631579 | 97 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/4-second-half-tasks/radosbench.yaml | meta:
- desc: |
run randomized correctness test for rados operations
generate write load with rados bench
second-half-tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90... | 362 | 20.352941 | 55 | yaml |
null | ceph-main/qa/suites/upgrade/pacific-x/stress-split/4-second-half-tasks/rbd-import-export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
second-half-tasks:
- workunit:
branch: pacific
clients:
client.0:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 5-workload"
| 270 | 19.846154 | 52 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/filestore-remove-check/2 - upgrade.yaml | meta:
- desc: |
install upgrade ceph/-x on cluster
restart : mons, osd.*
tasks:
- install.upgrade:
mon.a:
- exec:
osd.0:
- ceph osd require-osd-release quincy
- print: "**** done install.upgrade of nodes"
- ceph.restart:
daemons: [mon.a,mgr.x,osd.0,osd.1,osd.2]
mon-health-to-clog: false
... | 569 | 26.142857 | 80 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/filestore-remove-check/ubuntu_20.04.yaml | os_type: ubuntu
os_version: "20.04"
# the normal ubuntu 20.04 kernel (5.4.0-88-generic currently) have a bug that prevents the nvme_loop
# from behaving. I think it is this:
# https://lkml.org/lkml/2020/9/21/1456
# (at least, that is the symptom: nvme nvme1: Connect command failed, error wo/DNR bit: 880)
overrides:
... | 343 | 37.222222 | 100 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/filestore-remove-check/0-cluster/openstack.yaml | openstack:
- machine:
disk: 100 # GB
- volumes: # attached to each instance
count: 4
size: 30 # GB
| 121 | 16.428571 | 40 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/filestore-remove-check/0-cluster/start.yaml | meta:
- desc: |
Run ceph on one nodes,
Use xfs beneath the osds. upgrade to reef
should fail to start the osds with filestore
overrides:
ceph:
mon_bind_msgr2: false
mon_bind_addrvec: false
mon-health-to-clog: false
wait-for-healthy: false
wait-for-osds-up: false
wait-for-scrub: false
... | 770 | 22.363636 | 71 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/filestore-remove-check/1-ceph-install/quincy.yaml | meta:
- desc: install ceph/quincy latest
tasks:
- install:
exclude_packages:
- ceph-mgr-cephadm
- cephadm
- libcephfs-dev
branch: quincy
- print: "**** done install quincy"
- ceph:
create_rbd_pool: false
conf:
global:
bluestore_warn_on_legacy_statfs: false
bluesto... | 789 | 22.939394 | 61 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/filestore-remove-check/objectstore/filestore-xfs.yaml | overrides:
ceph:
fs: xfs
conf:
osd:
osd objectstore: filestore
osd sloppy crc: true
ceph-deploy:
fs: xfs
filestore: True
conf:
osd:
osd objectstore: filestore
osd sloppy crc: true | 247 | 16.714286 | 34 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/0-start.yaml | roles:
- - mon.a
- mon.c
- mgr.y
- mds.a
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- node-exporter.a
- alertmanager.a
- - mon.b
- mds.b
- mgr.x
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
- prometheus.a
- grafana.a
- node-exporter.b
openstack:
- volumes: # attached to each instan... | 460 | 12.558824 | 39 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/1-tasks.yaml | tasks:
- install:
branch: quincy
exclude_packages:
- ceph-volume
- print: "**** done install task..."
- print: "**** done start installing quincy cephadm ..."
- cephadm:
image: quay.ceph.io/ceph-ci/ceph:quincy
cephadm_branch: quincy
cephadm_git_url: https://github.com/ceph/ceph
conf:
... | 1,202 | 26.340909 | 80 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/upgrade-sequence.yaml | # renamed tasks: to upgrade-sequence:
upgrade-sequence:
sequential:
- print: "**** done start upgrade, wait"
- cephadm.shell:
env: [sha1]
mon.a:
- ceph config set global log_to_journald false --force
- ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1
- wh... | 650 | 37.294118 | 125 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/workload/ec-rados-default.yaml | meta:
- desc: |
run run randomized correctness test for rados operations
on an erasure-coded pool
workload:
full_sequential:
- print: "**** done start ec-rados-default.yaml"
- rados:
clients: [client.0]
ops: 4000
objects: 50
ec_pool: true
write_append_excl: false
op_weigh... | 577 | 21.230769 | 59 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/workload/rados_api.yaml | meta:
- desc: |
object class functional tests
workload:
full_sequential:
- print: "**** done start rados_api.yaml"
- workunit:
branch: quincy
clients:
client.0:
- cls
- print: "**** done end rados_api.yaml"
| 262 | 19.230769 | 45 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/workload/rados_loadgenbig.yaml | meta:
- desc: |
generate read/write load with rados objects ranging from 1MB to 25MB
workload:
full_sequential:
- print: "**** done start rados_loadgenbig.yaml"
- workunit:
branch: quincy
clients:
client.0:
- rados/load-gen-big.sh
- print: "**** done end rados_load... | 333 | 24.692308 | 71 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/workload/rbd_import_export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
workload:
full_sequential:
- print: "**** done start rbd_import_export.yaml"
- workunit:
branch: quincy
clients:
client.1:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- prin... | 362 | 23.2 | 53 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/workload/test_rbd_api.yaml | meta:
- desc: |
librbd C and C++ api tests
workload:
full_sequential:
- print: "**** done start test_rbd_api.yaml"
- workunit:
branch: quincy
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done end test_rbd_api.yaml"
| 282 | 20.769231 | 48 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/parallel/workload/test_rbd_python.yaml | meta:
- desc: |
librbd python api tests
workload:
full_sequential:
- print: "**** done start test_rbd_python.yaml"
- workunit:
branch: quincy
clients:
client.0:
- rbd/test_librbd_python.sh
- print: "**** done end test_rbd_python.yaml"
| 291 | 19.857143 | 51 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/0-roles.yaml | roles:
- - mon.a
- mon.c
- mgr.y
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- node-exporter.a
- alertmanager.a
- - mon.b
- mgr.x
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
- prometheus.a
- grafana.a
- node-exporter.b
openstack:
- volumes: # attached to each instance
count: 4
... | 440 | 12.78125 | 39 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/1-start.yaml | tasks:
- install:
branch: quincy
exclude_packages:
- ceph-volume
- cephadm:
image: quay.ceph.io/ceph-ci/ceph:quincy
cephadm_branch: quincy
cephadm_git_url: https://github.com/ceph/ceph
conf:
osd:
#set config option for which cls modules are allowed to be loaded / used
... | 3,185 | 24.902439 | 126 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/2-first-half-tasks/radosbench.yaml | meta:
- desc: |
run randomized correctness test for rados operations
generate write load with rados bench
first-half-tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
... | 418 | 19.95 | 55 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/2-first-half-tasks/rbd-cls.yaml | meta:
- desc: |
run basic cls tests for rbd
first-half-tasks:
- workunit:
branch: quincy
clients:
client.0:
- cls/test_cls_rbd.sh
- print: "**** done cls/test_cls_rbd.sh 5-workload"
| 207 | 17.909091 | 51 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/2-first-half-tasks/rbd-import-export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
first-half-tasks:
- workunit:
branch: quincy
clients:
client.0:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 5-workload"
| 268 | 19.692308 | 52 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/2-first-half-tasks/rbd_api.yaml | meta:
- desc: |
librbd C and C++ api tests
first-half-tasks:
- workunit:
branch: quincy
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done rbd/test_librbd.sh 7-workload"
| 211 | 18.272727 | 50 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/2-first-half-tasks/readwrite.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool,
using only reads, writes, and deletes
first-half-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 500
write_append_excl: false
op_weights:
read: 45
write:... | 391 | 22.058824 | 73 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/2-first-half-tasks/snaps-few-objects.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool with snapshot operations
first-half-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 50
write_append_excl: false
op_weights:
read: 100
write: 100
dele... | 452 | 22.842105 | 97 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/3-stress-tasks/radosbench.yaml | meta:
- desc: |
run randomized correctness test for rados operations
generate write load with rados bench
stress-tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- ... | 528 | 19.346154 | 55 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/3-stress-tasks/rbd-cls.yaml | meta:
- desc: |
run basic cls tests for rbd
stress-tasks:
- workunit:
branch: quincy
clients:
client.0:
- cls/test_cls_rbd.sh
- print: "**** done cls/test_cls_rbd.sh 5-workload"
| 203 | 17.545455 | 51 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/3-stress-tasks/rbd-import-export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
stress-tasks:
- workunit:
branch: quincy
clients:
client.0:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 5-workload"
| 264 | 19.384615 | 52 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/3-stress-tasks/rbd_api.yaml | meta:
- desc: |
librbd C and C++ api tests
stress-tasks:
- workunit:
branch: quincy
clients:
client.0:
- rbd/test_librbd.sh
- print: "**** done rbd/test_librbd.sh 7-workload"
| 207 | 17.909091 | 50 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/3-stress-tasks/readwrite.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool,
using only reads, writes, and deletes
stress-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 500
write_append_excl: false
op_weights:
read: 45
write: 45
... | 387 | 21.823529 | 73 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/3-stress-tasks/snaps-few-objects.yaml | meta:
- desc: |
randomized correctness test for rados operations on a replicated pool with snapshot operations
stress-tasks:
- full_sequential:
- rados:
clients: [client.0]
ops: 4000
objects: 50
write_append_excl: false
op_weights:
read: 100
write: 100
delete: ... | 448 | 22.631579 | 97 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/4-second-half-tasks/radosbench.yaml | meta:
- desc: |
run randomized correctness test for rados operations
generate write load with rados bench
second-half-tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90... | 362 | 20.352941 | 55 | yaml |
null | ceph-main/qa/suites/upgrade/quincy-x/stress-split/4-second-half-tasks/rbd-import-export.yaml | meta:
- desc: |
run basic import/export cli tests for rbd
second-half-tasks:
- workunit:
branch: quincy
clients:
client.0:
- rbd/import_export.sh
env:
RBD_CREATE_ARGS: --new-format
- print: "**** done rbd/import_export.sh 5-workload"
| 269 | 19.769231 | 52 | yaml |
null | ceph-main/qa/suites/upgrade/telemetry-upgrade/pacific-x/0-start.yaml | roles:
- - mon.a
- mon.c
- mgr.y
- mds.a
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- node-exporter.a
- alertmanager.a
- - mon.b
- mds.b
- mgr.x
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
- prometheus.a
- grafana.a
- node-exporter.b
openstack:
- volumes: # attached to each instan... | 460 | 12.558824 | 39 | yaml |
null | ceph-main/qa/suites/upgrade/telemetry-upgrade/pacific-x/1-tasks.yaml | tasks:
- install:
branch: pacific
exclude_packages:
- ceph-volume
- print: "**** done install task..."
- print: "**** done start installing pacific cephadm ..."
- cephadm:
image: quay.io/ceph/daemon-base:latest-pacific
cephadm_branch: pacific
cephadm_git_url: https://github.com/ceph/ceph
c... | 1,823 | 32.163636 | 128 | yaml |
null | ceph-main/qa/suites/upgrade/telemetry-upgrade/quincy-x/0-start.yaml | roles:
- - mon.a
- mon.c
- mgr.y
- mds.a
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- node-exporter.a
- alertmanager.a
- - mon.b
- mds.b
- mgr.x
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
- prometheus.a
- grafana.a
- node-exporter.b
openstack:
- volumes: # attached to each instan... | 460 | 12.558824 | 39 | yaml |
null | ceph-main/qa/suites/upgrade/telemetry-upgrade/quincy-x/1-tasks.yaml | tasks:
- install:
branch: quincy
exclude_packages:
- ceph-volume
- print: "**** done install task..."
- print: "**** done start installing quincy cephadm ..."
- cephadm:
image: quay.io/ceph/daemon-base:latest-quincy
cephadm_branch: quincy
cephadm_git_url: https://github.com/ceph/ceph
conf:... | 1,813 | 32.592593 | 128 | yaml |
null | ceph-main/qa/suites/windows/basic/ubuntu_latest.yaml | .qa/distros/supported/ubuntu_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/windows/basic/clusters/fixed-1.yaml | .qa/clusters/fixed-1.yaml | 25 | 25 | 25 | yaml |
null | ceph-main/qa/suites/windows/basic/install/install.yaml | tasks:
- install:
- ceph:
| 26 | 5.75 | 10 | yaml |
null | ceph-main/qa/suites/windows/basic/tasks/windows_tests.yaml | tasks:
- workunit:
clients:
client.0:
- windows/libvirt_vm/setup.sh
- windows/run-tests.sh
| 117 | 15.857143 | 37 | yaml |
null | ceph-main/qa/tasks/__init__.py | import logging
# Inherit teuthology's log level
teuthology_log = logging.getLogger('teuthology')
log = logging.getLogger(__name__)
log.setLevel(teuthology_log.level)
| 167 | 23 | 48 | py |
null | ceph-main/qa/tasks/admin_socket.py | """
Admin Socket task -- used in rados, powercycle, and smoke testing
"""
import json
import logging
import os
import time
from teuthology.exceptions import CommandFailedError
from teuthology.orchestra import run
from teuthology import misc as teuthology
from teuthology.parallel import parallel
from teuthology.config... | 6,476 | 30.595122 | 79 | py |
null | ceph-main/qa/tasks/autotest.py | """
Run an autotest test on the ceph cluster.
"""
import json
import logging
import os
from teuthology import misc as teuthology
from teuthology.parallel import parallel
from teuthology.orchestra import run
log = logging.getLogger(__name__)
def task(ctx, config):
"""
Run an autotest test on the ceph cluster.... | 4,983 | 29.024096 | 87 | py |
null | ceph-main/qa/tasks/aver.py | """
Aver wrapper task
"""
import contextlib
import logging
from subprocess import check_call, Popen, PIPE
log = logging.getLogger(__name__)
@contextlib.contextmanager
def task(ctx, config):
"""
Execute an aver assertion
Parameters:
input: file containing data referred to by the assertions. File... | 2,064 | 29.367647 | 79 | py |
null | ceph-main/qa/tasks/backfill_toofull.py | """
Backfill_toofull
"""
import logging
import time
from tasks import ceph_manager
from tasks.util.rados import rados
from teuthology import misc as teuthology
log = logging.getLogger(__name__)
def wait_for_pg_state(manager, pgid, state, to_osd):
log.debug("waiting for pg %s state is %s" % (pgid, state))
for ... | 6,634 | 33.201031 | 85 | py |
null | ceph-main/qa/tasks/barbican.py | """
Deploy and configure Barbican for Teuthology
"""
import argparse
import contextlib
import logging
import http
import json
import time
import math
from urllib.parse import urlparse
from teuthology import misc as teuthology
from teuthology import contextutil
from teuthology.orchestra import run
from teuthology.exce... | 19,518 | 36.108365 | 102 | py |
null | ceph-main/qa/tasks/blktrace.py | """
Run blktrace program through teuthology
"""
import contextlib
import logging
from teuthology import misc as teuthology
from teuthology import contextutil
from teuthology.orchestra import run
log = logging.getLogger(__name__)
blktrace = '/usr/sbin/blktrace'
daemon_signal = 'term'
@contextlib.contextmanager
def se... | 2,821 | 28.092784 | 92 | py |
null | ceph-main/qa/tasks/cbt.py | import logging
import os
import yaml
from teuthology import misc
from teuthology.orchestra import run
from teuthology.task import Task
log = logging.getLogger(__name__)
class CBT(Task):
"""
Passes through a CBT configuration yaml fragment.
"""
def __init__(self, ctx, config):
super(CBT, self... | 12,249 | 39.03268 | 143 | py |
null | ceph-main/qa/tasks/ceph.py | """
Ceph cluster task.
Handle the setup, starting, and clean-up of a Ceph cluster.
"""
from copy import deepcopy
from io import BytesIO
from io import StringIO
import argparse
import configobj
import contextlib
import errno
import logging
import os
import json
import time
import gevent
import re
import socket
import ... | 68,221 | 33.683274 | 113 | py |
null | ceph-main/qa/tasks/ceph_client.py | """
Set up client keyring
"""
import logging
from teuthology import misc as teuthology
from teuthology.orchestra import run
log = logging.getLogger(__name__)
def create_keyring(ctx, cluster_name):
"""
Set up key ring on remote sites
"""
log.info('Setting up client nodes...')
clients = ctx.cluster... | 1,501 | 33.930233 | 104 | py |
null | ceph-main/qa/tasks/ceph_deploy.py | """
Execute ceph-deploy as a task
"""
import contextlib
import os
import time
import logging
import traceback
from teuthology import misc as teuthology
from teuthology import contextutil
from teuthology.config import config as teuth_config
from teuthology.task import install as install_fn
from teuthology.orchestra im... | 34,996 | 36.916576 | 90 | py |
null | ceph-main/qa/tasks/ceph_fuse.py | """
Ceph FUSE client task
"""
import contextlib
import logging
from teuthology import misc
from tasks.cephfs.fuse_mount import FuseMount
log = logging.getLogger(__name__)
@contextlib.contextmanager
def task(ctx, config):
"""
Mount/unmount a ``ceph-fuse`` client.
The config is optional and defaults to ... | 5,946 | 30.973118 | 109 | py |