Search is not available for this dataset
repo stringlengths 2 152 ⌀ | file stringlengths 15 239 | code stringlengths 0 58.4M | file_length int64 0 58.4M | avg_line_length float64 0 1.81M | max_line_length int64 0 12.7M | extension_type stringclasses 364
values |
|---|---|---|---|---|---|---|
null | ceph-main/qa/suites/orch/rook/smoke/2-workload/none.yaml | 0 | 0 | 0 | yaml | |
null | ceph-main/qa/suites/orch/rook/smoke/2-workload/radosbench.yaml | tasks:
- install:
host.a:
- radosbench:
clients: [client.a]
| 68 | 10.5 | 23 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/cluster/1-node.yaml | overrides:
ceph:
conf:
global:
osd crush chooseleaf type: 0
roles:
- - host.a
- client.a
| 112 | 10.3 | 36 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/cluster/3-node.yaml | roles:
- - host.a
- client.a
- - host.b
- client.b
- - host.c
- client.c
| 79 | 9 | 12 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/k8s/1.21.yaml | overrides:
kubeadm:
version: "1.21"
| 42 | 9.75 | 19 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/net/calico.yaml | overrides:
kubeadm:
pod_network: calico
| 46 | 10.75 | 23 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/net/flannel.yaml | overrides:
kubeadm:
pod_network: flannel
| 47 | 11 | 24 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/net/host.yaml | overrides:
ceph:
spec:
mon:
allowMultiplePerNode: false
network:
provider: host
| 114 | 13.375 | 35 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/rook/1.7.2.yaml | overrides:
rook:
rook_image: rook/ceph:v1.7.2
rook_branch: v1.7.2
| 76 | 14.4 | 32 | yaml |
null | ceph-main/qa/suites/orch/rook/smoke/rook/master.yaml | overrides:
rook:
rook_image: rook/ceph:master
| 52 | 12.25 | 32 | yaml |
null | ceph-main/qa/suites/perf-basic/ceph.yaml | meta:
- desc: |
perf-basic is a basic performance suite.
Must be run on bare-metal machines.
On VMs performance results will be inconsistent
and can't be compared across runs.
Run ceph on a single node.
Use xfs beneath the osds.
Setup rgw on client.0
roles:
- [mon.a, mgr.x, osd.0, osd.1, osd.2, cl... | 494 | 19.625 | 50 | yaml |
null | ceph-main/qa/suites/perf-basic/ubuntu_latest.yaml | .qa/distros/supported/ubuntu_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/perf-basic/objectstore/bluestore.yaml | overrides:
ceph:
fs: xfs
conf:
osd:
osd objectstore: bluestore
bluestore block size: 96636764160
ceph-deploy:
fs: xfs
bluestore: yes
conf:
osd:
osd objectstore: bluestore
bluestore block size: 96636764160
| 274 | 16.1875 | 41 | yaml |
null | ceph-main/qa/suites/perf-basic/settings/optimized.yaml | meta:
- desc: |
Use debug level 0/0 for performance tests.
overrides:
ceph:
conf:
mon:
debug mon: "0/0"
debug ms: "0/0"
debug paxos: "0/0"
osd:
debug filestore: "0/0"
debug journal: "0/0"
debug ms: "0/0"
debug osd: "0/0"
global:
... | 2,033 | 25.076923 | 45 | yaml |
null | ceph-main/qa/suites/perf-basic/workloads/client_endpoint_rbd_4K_rand_write.yaml | meta:
- desc: |
Run librbdfio benchmark using cbt client endpoint for rbd.
4K randwrite workload.
tasks:
- cbt:
benchmarks:
fio:
client_endpoints: 'fiotest'
op_size: [4096]
time: 300
mode: ['randwrite']
norandommap: True
size: 4096
iodepth: [32]
... | 642 | 18.484848 | 61 | yaml |
null | ceph-main/qa/suites/perf-basic/workloads/fio_4K_rand_write.yaml | meta:
- desc: |
Run librbdfio benchmark using cbt.
4K randwrite workload.
tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4096]
time: 300
mode: ['randwrite']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iod... | 591 | 18.733333 | 37 | yaml |
null | ceph-main/qa/suites/perf-basic/workloads/radosbench_4K_write.yaml | meta:
- desc: |
Run radosbench benchmark using cbt.
4K write workload.
tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4096]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
... | 580 | 19.034483 | 38 | yaml |
null | ceph-main/qa/suites/powercycle/osd/ignorelist_health.yaml | overrides:
ceph:
log-ignorelist:
- \(MDS_TRIM\)
- \(MDS_SLOW_REQUEST\)
- MDS_SLOW_METADATA_IO
- Behind on trimming
| 145 | 17.25 | 28 | yaml |
null | ceph-main/qa/suites/powercycle/osd/thrashosds-health.yaml | .qa/tasks/thrashosds-health.yaml | 32 | 32 | 32 | yaml |
null | ceph-main/qa/suites/powercycle/osd/clusters/3osd-1per-target.yaml | roles:
- [mon.a, mon.b, mon.c, mgr.x, mgr.y, mds.a, client.0]
- [osd.0]
- [osd.1]
- [osd.2]
| 92 | 14.5 | 54 | yaml |
null | ceph-main/qa/suites/powercycle/osd/powercycle/default.yaml | tasks:
- install:
extra_system_packages:
deb:
- bison
- flex
- libelf-dev
- libssl-dev
- libaio-dev
- libtool-bin
- uuid-dev
- xfslibs-dev
rpm:
- bison
- flex
- elfutils-libelf-devel
- openssl-devel
- libaio-devel
- libt... | 450 | 15.703704 | 29 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/admin_socket_objecter_requests.yaml | overrides:
ceph:
conf:
client.0:
admin socket: /var/run/ceph/ceph-$name.asok
tasks:
- radosbench:
clients: [client.0]
time: 60
- admin_socket:
client.0:
objecter_requests:
test: "http://git.ceph.com/?p={repo};a=blob_plain;f=src/test/admin_socket/objecter_requests;hb={branch... | 323 | 22.142857 | 112 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_kernel_untar_build.yaml | overrides:
ceph:
conf:
client:
fuse_default_permissions: 0
tasks:
- ceph-fuse:
- workunit:
timeout: 6h
clients:
all:
- kernel_untar_build.sh
| 183 | 13.153846 | 35 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_misc.yaml | tasks:
- ceph-fuse:
- workunit:
timeout: 6h
clients:
all:
- fs/misc
| 90 | 10.375 | 17 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_suites_ffsb.yaml | overrides:
ceph:
conf:
osd:
filestore flush min: 0
mds:
debug ms: 1
debug mds: 20
tasks:
- ceph-fuse:
- workunit:
clients:
all:
- suites/ffsb.sh
| 205 | 12.733333 | 30 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_suites_fsstress.yaml | tasks:
- ceph-fuse:
- workunit:
clients:
all:
- suites/fsstress.sh
| 85 | 11.285714 | 28 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_suites_fsx.yaml | tasks:
- ceph-fuse:
- workunit:
timeout: 6h
clients:
all:
- suites/fsx.sh
| 96 | 11.125 | 23 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_suites_fsync.yaml | overrides:
ceph:
conf:
global:
osd_pg_log_dups_tracked: 10000
tasks:
- ceph-fuse:
- workunit:
clients:
all:
- suites/fsync-tester.sh
| 172 | 12.307692 | 38 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_suites_pjd.yaml | overrides:
ceph:
conf:
client:
fuse set user groups: true
tasks:
- ceph-fuse:
- workunit:
clients:
all:
- suites/pjd.sh
| 158 | 12.25 | 34 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/cfuse_workunit_suites_truncate_delay.yaml | overrides:
ceph:
conf:
client:
ms_inject_delay_probability: 1
ms_inject_delay_type: osd
ms_inject_delay_max: 5
client_oc_max_dirty_age: 1
tasks:
- ceph-fuse:
- exec:
client.0:
- dd if=/dev/zero of=./foo count=100
- sleep 2
- truncate --size 0 ./foo
| 315 | 18.75 | 42 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- \(POOL_APP_NOT_ENABLED\)
- \(PG_AVAILABILITY\)
conf:
mon:
mon warn on pool no app: false
osd:
osd class load list: "*"
osd class default list: "*"
tasks:
- ceph-fuse:
- workunit:
clients:
client.... | 347 | 17.315789 | 38 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/radosbench.yaml | tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- rados... | 710 | 17.230769 | 25 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/readwrite.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
op_weights:
read: 45
write: 45
delete: 10
| 135 | 12.6 | 23 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/snaps-few-objects.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 219 | 14.714286 | 23 | yaml |
null | ceph-main/qa/suites/powercycle/osd/tasks/snaps-many-objects.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 220 | 14.785714 | 23 | yaml |
null | ceph-main/qa/suites/rados/basic/ceph.yaml | overrides:
ceph:
conf:
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
tasks:
- install:
extra_system_packages:
r... | 380 | 19.052632 | 41 | yaml |
null | ceph-main/qa/suites/rados/basic/rados.yaml | .qa/config/rados.yaml | 21 | 21 | 21 | yaml |
null | ceph-main/qa/suites/rados/basic/clusters/fixed-2.yaml | .qa/clusters/fixed-2.yaml | 25 | 25 | 25 | yaml |
null | ceph-main/qa/suites/rados/basic/clusters/openstack.yaml | openstack:
- volumes: # attached to each instance
count: 4
size: 10 # GB
| 87 | 16.6 | 40 | yaml |
null | ceph-main/qa/suites/rados/basic/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rados/basic/msgr-failures/many.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 1500
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/libcephsqlite.yaml | overrides:
ceph:
conf:
client:
debug ms: 1
debug client: 20
debug cephsqlite: 20
log-ignorelist:
- POOL_APP_NOT_ENABLED
- do not have an application enabled
tasks:
- exec:
client.0:
- ceph osd pool create cephsqlite
- ceph auth get-or-create client.libcephsqlite ... | 662 | 25.52 | 156 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- but it is still running
- overall HEALTH_
- \(POOL_FULL\)
- \(SMALLER_PGP_NUM\)
- \(CACHE_POOL_NO_HIT_SET\)
- \(CACHE_POOL_NEAR_FULL\)
- \(POOL_APP_NOT_ENABLED\)
- \(PG_AVAILABILITY\)
- \(PG_DEGRADED\)
conf:
clien... | 581 | 19.068966 | 38 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_cls_all.yaml | overrides:
ceph:
log-ignorelist:
- \(PG_AVAILABILITY\)
- \(POOL_APP_NOT_ENABLED\)
conf:
osd:
osd_class_load_list: "*"
osd_class_default_list: "*"
tasks:
- workunit:
clients:
client.0:
- cls
| 248 | 15.6 | 35 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_python.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- overall HEALTH_
- \(OSDMAP_FLAGS\)
- \(PG_
- \(OSD_
- \(OBJECT_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- workunit:
timeout: 1h
clients:
client.0:
- rados/test_python.sh
| 281 | 15.588235 | 30 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_stress_watch.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
- \(TOO_FEW_PGS\)
tasks:
- workunit:
clients:
client.0:
- rados/stress_watch.sh
| 201 | 15.833333 | 33 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_striper.yaml | tasks:
- exec:
client.0:
- ceph_test_rados_striper_api_io
- ceph_test_rados_striper_api_aio
- ceph_test_rados_striper_api_striping
| 144 | 17.125 | 41 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_workunit_loadgen_big.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- overall HEALTH_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- workunit:
clients:
all:
- rados/load-gen-big.sh
| 197 | 15.5 | 31 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mix.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- overall HEALTH_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- workunit:
clients:
all:
- rados/load-gen-mix.sh
| 197 | 15.5 | 31 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mostlyread.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- overall HEALTH_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- workunit:
clients:
all:
- rados/load-gen-mostlyread.sh
| 204 | 16.083333 | 38 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/readwrite.yaml | overrides:
ceph:
crush_tunables: optimal
conf:
mon:
mon osd initial require min compat client: luminous
osd:
osd_discard_disconnected_ops: false
tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
op_weights:
read: 45
write: 45
delete: 10
| 318 | 16.722222 | 59 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/repair_test.yaml | overrides:
ceph:
wait-for-scrub: false
log-ignorelist:
- candidate had a stat error
- candidate had a read error
- deep-scrub 0 missing, 1 inconsistent objects
- deep-scrub 0 missing, 4 inconsistent objects
- deep-scrub [0-9]+ errors
- '!= omap_digest'
- '!= data_dige... | 926 | 27.96875 | 62 | yaml |
null | ceph-main/qa/suites/rados/basic/tasks/scrub_test.yaml | overrides:
ceph:
wait-for-scrub: false
log-ignorelist:
- '!= data_digest'
- '!= omap_digest'
- '!= size'
- 'deep-scrub 0 missing, 1 inconsistent objects'
- 'deep-scrub [0-9]+ errors'
- 'repair 0 missing, 1 inconsistent objects'
- 'repair [0-9]+ errors, [0-9]+ fixed'
- 'shard [0... | 858 | 26.709677 | 55 | yaml |
null | ceph-main/qa/suites/rados/dashboard/0-single-container-host.yaml | .qa/distros/single-container-host.yaml | 38 | 38 | 38 | yaml |
null | ceph-main/qa/suites/rados/dashboard/debug/mgr.yaml | .qa/debug/mgr.yaml | 18 | 18 | 18 | yaml |
null | ceph-main/qa/suites/rados/dashboard/tasks/dashboard.yaml | roles:
- [mgr.x, mon.a, mon.c, mds.a, mds.c, osd.0, client.0]
- [mgr.y, mgr.z, mon.b, mds.b, osd.1, osd.2, osd.3, client.1]
overrides:
ceph:
conf:
osd:
osd mclock override recovery settings: true
tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
... | 2,356 | 32.197183 | 79 | yaml |
null | ceph-main/qa/suites/rados/dashboard/tasks/e2e.yaml | roles:
# 3 osd roles on host.a is required for cephadm task. It checks if the cluster is healthy.
# More daemons will be deployed on both hosts in e2e tests.
- - host.a
- osd.0
- osd.1
- osd.2
- mon.a
- mgr.a
- client.0
- - host.b
- client.1
tasks:
- install:
- cephadm:
- workunit:
clients:
clie... | 447 | 17.666667 | 90 | yaml |
null | ceph-main/qa/suites/rados/mgr/clusters/2-node-mgr.yaml | .qa/clusters/2-node-mgr.yaml | 28 | 28 | 28 | yaml |
null | ceph-main/qa/suites/rados/mgr/debug/mgr.yaml | .qa/debug/mgr.yaml | 18 | 18 | 18 | yaml |
null | ceph-main/qa/suites/rados/mgr/mgr_ttl_cache/disable.yaml | .qa/mgr_ttl_cache/disable.yaml | 30 | 30 | 30 | yaml |
null | ceph-main/qa/suites/rados/mgr/mgr_ttl_cache/enable.yaml | .qa/mgr_ttl_cache/enable.yaml | 29 | 29 | 29 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/crash.yaml |
tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dump during teardown.
wait-for-scrub: false
log-ignorelist:
- overall HEALTH_
- \(MGR_DOWN\)
- \(PG_
- \(RECENT_CRASH\)
- replacing it with st... | 435 | 23.222222 | 68 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/failover.yaml |
tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dump during teardown.
wait-for-scrub: false
log-ignorelist:
- overall HEALTH_
- \(MGR_DOWN\)
- \(PG_
- replacing it with standby
- No standby ... | 411 | 23.235294 | 68 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/insights.yaml |
tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dump during teardown.
wait-for-scrub: false
log-ignorelist:
- overall HEALTH_
- \(MGR_DOWN\)
- \(MGR_INSIGHTS_WARNING\)
- \(insights_health_check
... | 507 | 24.4 | 68 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/module_selftest.yaml |
tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dump during teardown.
wait-for-scrub: false
log-ignorelist:
- overall HEALTH_
- \(MGR_DOWN\)
- \(PG_
- replacing it with standby
- No standby ... | 814 | 28.107143 | 73 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/per_module_finisher_stats.yaml | tasks:
- install:
- ceph:
wait-for-scrub: false
- check-counter:
counters:
mgr:
- name: "finisher-balancer.complete_latency.avgcount"
min: 1
- name: "finisher-balancer.queue_len"
expected_val: 0
- name: "finisher-crash.complete_la... | 1,539 | 34 | 70 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/progress.yaml | overrides:
ceph:
conf:
osd:
osd mclock profile: high_recovery_ops
tasks:
- install:
- ceph:
config:
global:
osd pool default size : 3
osd pool default min size : 2
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dum... | 750 | 24.033333 | 68 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/prometheus.yaml |
tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dump during teardown.
wait-for-scrub: false
log-ignorelist:
- overall HEALTH_
- \(MGR_DOWN\)
- \(PG_
- replacing it with standby
- No standby ... | 413 | 23.352941 | 68 | yaml |
null | ceph-main/qa/suites/rados/mgr/tasks/workunits.yaml | tasks:
- install:
- ceph:
# tests may leave mgrs broken, so don't try and call into them
# to invoke e.g. pg dump during teardown.
wait-for-scrub: false
log-ignorelist:
- overall HEALTH_
- \(MGR_DOWN\)
- \(PG_
- replacing it with standby
- No standby d... | 418 | 23.647059 | 68 | yaml |
null | ceph-main/qa/suites/rados/monthrash/ceph.yaml | overrides:
ceph:
conf:
client:
debug monc: 20
debug ms: 1
mon:
mon min osdmap epochs: 25
paxos service trim min: 5
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize... | 727 | 24.103448 | 63 | yaml |
null | ceph-main/qa/suites/rados/monthrash/rados.yaml | .qa/config/rados.yaml | 21 | 21 | 21 | yaml |
null | ceph-main/qa/suites/rados/monthrash/clusters/3-mons.yaml | roles:
- [mon.a, mon.c, osd.0, osd.1, osd.2]
- [mon.b, mgr.x, osd.3, osd.4, osd.5, client.0]
openstack:
- volumes: # attached to each instance
count: 3
size: 10 # GB
| 180 | 21.625 | 47 | yaml |
null | ceph-main/qa/suites/rados/monthrash/clusters/9-mons.yaml | roles:
- [mon.a, mon.b, mon.c, mon.d, mon.e, osd.0, osd.1, osd.2]
- [mon.f, mon.g, mon.h, mon.i, mgr.x, osd.3, osd.4, osd.5, client.0]
openstack:
- volumes: # attached to each instance
count: 3
size: 10 # GB
| 222 | 26.875 | 68 | yaml |
null | ceph-main/qa/suites/rados/monthrash/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rados/monthrash/msgr-failures/mon-delay.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 2500
ms inject delay type: mon
ms inject delay probability: .005
ms inject delay max: 1
ms inject internal delays: .002
mon client directed command retry: 5
mgr:
debug monc: 10
log-ign... | 358 | 22.933333 | 44 | yaml |
null | ceph-main/qa/suites/rados/monthrash/thrashers/force-sync-many.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
- \(TOO_FEW_PGS\)
tasks:
- mon_thrash:
revive_delay: 90
thrash_delay: 1
store_thrash: true
thrash_many: true
| 215 | 15.615385 | 23 | yaml |
null | ceph-main/qa/suites/rados/monthrash/thrashers/many.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
conf:
osd:
mon client ping interval: 4
mon client ping timeout: 12
tasks:
- mon_thrash:
revive_delay: 20
thrash_delay: 1
thrash_many: true
freeze_mon_duration: 20
freeze_mon_probability: 10
| 320 | 17.882353 | 35 | yaml |
null | ceph-main/qa/suites/rados/monthrash/thrashers/one.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
tasks:
- mon_thrash:
revive_delay: 20
thrash_delay: 1
| 146 | 13.7 | 23 | yaml |
null | ceph-main/qa/suites/rados/monthrash/thrashers/sync-many.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
conf:
mon:
paxos min: 10
paxos trim min: 10
tasks:
- mon_thrash:
revive_delay: 90
thrash_delay: 1
thrash_many: true
| 238 | 14.933333 | 26 | yaml |
null | ceph-main/qa/suites/rados/monthrash/thrashers/sync.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
conf:
mon:
paxos min: 10
paxos trim min: 10
tasks:
- mon_thrash:
revive_delay: 90
thrash_delay: 1
| 216 | 14.5 | 26 | yaml |
null | ceph-main/qa/suites/rados/monthrash/workloads/pool-create-delete.yaml | overrides:
ceph:
log-ignorelist:
- slow request
- overall HEALTH_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- exec:
client.0:
- ceph_test_rados_delete_pools_parallel
- ceph_test_rados_delete_pools_parallel
- ceph_test_rados_delete_pools_parallel
- ceph_test_rados_delete_pools... | 2,400 | 39.694915 | 45 | yaml |
null | ceph-main/qa/suites/rados/monthrash/workloads/rados_5925.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- exec:
client.0:
- ceph_test_rados_delete_pools_parallel --debug_objecter 20 --debug_ms 1 --debug_rados 20 --debug_monc 20
| 240 | 23.1 | 111 | yaml |
null | ceph-main/qa/suites/rados/monthrash/workloads/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
- \(CACHE_POOL_NEAR_FULL\)
- \(POOL_FULL\)
- \(SLOW_OPS\)
- \(MON_DOWN\)
- \(PG_
- \(POOL_APP_NOT_ENABLED\)
- \(SMALLER_PGP_NUM\)
- slow request
conf... | 609 | 19.333333 | 38 | yaml |
null | ceph-main/qa/suites/rados/monthrash/workloads/rados_mon_osdmap_prune.yaml | overrides:
ceph:
conf:
mon:
mon debug extra checks: true
mon min osdmap epochs: 100
mon osdmap full prune enabled: true
mon osdmap full prune min: 200
mon osdmap full prune interval: 10
mon osdmap full prune txsize: 100
osd:
osd beacon report int... | 624 | 26.173913 | 58 | yaml |
null | ceph-main/qa/suites/rados/monthrash/workloads/rados_mon_workunits.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- overall HEALTH_
- \(PG_
- \(MON_DOWN\)
- \(AUTH_BAD_CAPS\)
tasks:
- workunit:
clients:
client.0:
- mon/pool_ops.sh
- mon/crush_ops.sh
- mon/osd.sh
- mon/caps.sh
| 291 | 15.222222 | 29 | yaml |
null | ceph-main/qa/suites/rados/monthrash/workloads/snaps-few-objects.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 219 | 14.714286 | 23 | yaml |
null | ceph-main/qa/suites/rados/multimon/no_pools.yaml | overrides:
ceph:
create_rbd_pool: false
pre-mgr-commands:
- sudo ceph config set mgr mgr_pool false --force
| 124 | 19.833333 | 55 | yaml |
null | ceph-main/qa/suites/rados/multimon/rados.yaml | .qa/config/rados.yaml | 21 | 21 | 21 | yaml |
null | ceph-main/qa/suites/rados/multimon/clusters/21.yaml | roles:
- [mon.a, mon.d, mon.g, mon.j, mon.m, mon.p, mon.s]
- [mon.b, mon.e, mon.h, mon.k, mon.n, mon.q, mon.t, mgr.x]
- [mon.c, mon.f, mon.i, mon.l, mon.o, mon.r, mon.u]
openstack:
- volumes: # attached to each instance
count: 1
size: 10 # GB
| 251 | 27 | 58 | yaml |
null | ceph-main/qa/suites/rados/multimon/clusters/3.yaml | roles:
- [mon.a, mon.c]
- [mon.b, mgr.x]
openstack:
- volumes: # attached to each instance
count: 2
size: 10 # GB
| 122 | 14.375 | 38 | yaml |
null | ceph-main/qa/suites/rados/multimon/clusters/6.yaml | roles:
- [mon.a, mon.c, mon.e, mgr.x]
- [mon.b, mon.d, mon.f, mgr.y]
openstack:
- volumes: # attached to each instance
count: 1
size: 10 # GB
| 150 | 17.875 | 38 | yaml |
null | ceph-main/qa/suites/rados/multimon/clusters/9.yaml | roles:
- [mon.a, mon.d, mon.g]
- [mon.b, mon.e, mon.h, mgr.x]
- [mon.c, mon.f, mon.i]
openstack:
- volumes: # attached to each instance
count: 1
size: 10 # GB
| 167 | 17.666667 | 38 | yaml |
null | ceph-main/qa/suites/rados/multimon/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rados/multimon/msgr-failures/many.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 1000
mon client directed command retry: 5
mon mgr beacon grace: 90
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 210 | 20.1 | 44 | yaml |
null | ceph-main/qa/suites/rados/multimon/tasks/mon_clock_no_skews.yaml | tasks:
- install:
- ceph:
log-ignorelist:
- slow request
- .*clock.*skew.*
- clocks not synchronized
- overall HEALTH_
- \(MON_CLOCK_SKEW\)
- mon_clock_skew_check:
expect-skew: false
| 211 | 16.666667 | 29 | yaml |
null | ceph-main/qa/suites/rados/multimon/tasks/mon_clock_with_skews.yaml | tasks:
- install:
- exec:
mon.b:
- sudo systemctl stop chronyd.service || true
- sudo systemctl stop systemd-timesync.service || true
- sudo systemctl stop ntpd.service || true
- sudo systemctl stop ntp.service || true
- date -u -s @$(expr $(date -u +%s) + 2)
- ceph:
wait-for-healthy: false
... | 608 | 23.36 | 58 | yaml |
null | ceph-main/qa/suites/rados/multimon/tasks/mon_recovery.yaml | tasks:
- install:
- ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
- \(PG_AVAILABILITY\)
- \(SLOW_OPS\)
- slow request
- mon_recovery:
| 177 | 15.181818 | 27 | yaml |
null | ceph-main/qa/suites/rados/objectstore/backends/ceph_objectstore_tool.yaml | roles:
- [mon.a, mgr.x, osd.0, osd.1, osd.2, osd.3, osd.4, osd.5, client.0]
openstack:
- volumes: # attached to each instance
count: 6
size: 10 # GB
tasks:
- install:
- ceph:
fs: xfs
conf:
global:
osd max object name len: 460
osd max object namespace len: 64
osd:
osd ... | 565 | 19.962963 | 68 | yaml |
null | ceph-main/qa/suites/rados/objectstore/backends/fusestore.yaml | roles:
- [mon.a, mgr.x, osd.0, osd.1, client.0]
tasks:
- install:
- workunit:
clients:
all:
- objectstore/test_fuse.sh
| 138 | 12.9 | 40 | yaml |
null | ceph-main/qa/suites/rados/objectstore/backends/keyvaluedb.yaml | roles:
- [mon.a, mgr.x, osd.0, osd.1, client.0]
tasks:
- install:
- exec:
client.0:
- mkdir $TESTDIR/kvtest && cd $TESTDIR/kvtest && ceph_test_keyvaluedb
- rm -rf $TESTDIR/kvtest
| 195 | 20.777778 | 75 | yaml |