Search is not available for this dataset
repo
stringlengths
2
152
file
stringlengths
15
239
code
stringlengths
0
58.4M
file_length
int64
0
58.4M
avg_line_length
float64
0
1.81M
max_line_length
int64
0
12.7M
extension_type
stringclasses
364 values
null
ceph-main/qa/suites/rados/thrash/2-recovery-overrides/default.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-active-recovery.yaml
.qa/overrides/more-active-recovery.yaml
39
39
39
yaml
null
ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-async-partial-recovery.yaml
overrides: ceph: conf: global: osd_async_recovery_min_cost: 1 osd_object_clean_region_max_num_intervals: 1000
138
18.857143
55
yaml
null
ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-async-recovery.yaml
overrides: ceph: conf: global: osd_async_recovery_min_cost: 1
82
12.833333
38
yaml
null
ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-partial-recovery.yaml
overrides: ceph: conf: global: osd_object_clean_region_max_num_intervals: 1000
99
15.666667
55
yaml
null
ceph-main/qa/suites/rados/thrash/3-scrub-overrides/default.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-2.yaml
overrides: ceph: conf: osd: osd max scrubs: 2
66
10.166667
25
yaml
null
ceph-main/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-3.yaml
overrides: ceph: conf: osd: osd max scrubs: 3
66
10.166667
25
yaml
null
ceph-main/qa/suites/rados/thrash/backoff/normal.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/thrash/backoff/peering.yaml
overrides: ceph: conf: osd: osd backoff on peering: true
77
12
36
yaml
null
ceph-main/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml
overrides: ceph: conf: osd: osd backoff on peering: true osd backoff on degraded: true
115
15.571429
37
yaml
null
ceph-main/qa/suites/rados/thrash/clusters/fixed-2.yaml
.qa/clusters/fixed-2.yaml
25
25
25
yaml
null
ceph-main/qa/suites/rados/thrash/clusters/openstack.yaml
openstack: - volumes: # attached to each instance count: 4 size: 30 # GB
87
16.6
40
yaml
null
ceph-main/qa/suites/rados/thrash/crc-failures/bad_map_crc_failure.yaml
overrides: ceph: conf: osd: osd inject bad map crc probability: 0.1 log-ignorelist: - failed to encode map
137
16.25
47
yaml
null
ceph-main/qa/suites/rados/thrash/crc-failures/default.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/thrash/d-balancer/crush-compat.yaml
tasks: - exec: mon.a: - while ! ceph balancer status ; do sleep 1 ; done - ceph balancer mode crush-compat - ceph balancer on
148
20.285714
56
yaml
null
ceph-main/qa/suites/rados/thrash/d-balancer/on.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/thrash/msgr-failures/fastclose.yaml
overrides: ceph: conf: global: ms inject socket failures: 2500 ms tcp read timeout: 5 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
208
19.9
44
yaml
null
ceph-main/qa/suites/rados/thrash/msgr-failures/few.yaml
overrides: ceph: conf: global: ms inject socket failures: 5000 mon client directed command retry: 5 osd: osd heartbeat use min delay socket: true log-ignorelist: - \(OSD_SLOW_PING_TIME
237
20.636364
48
yaml
null
ceph-main/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml
overrides: ceph: conf: global: ms inject socket failures: 2500 ms inject delay type: osd ms inject delay probability: .005 ms inject delay max: 1 ms inject internal delays: .002 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_T...
324
24
44
yaml
null
ceph-main/qa/suites/rados/thrash/msgr-failures/osd-dispatch-delay.yaml
overrides: ceph: conf: global: osd debug inject dispatch delay duration: 0.1 osd debug inject dispatch delay probability: 0.1
155
18.5
56
yaml
null
ceph-main/qa/suites/rados/thrash/thrashers/careful.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost conf: osd: osd debug reject backfill probability: .3 osd scrub min interval: 60 osd scrub max interval: 120 osd max backfills: 3 osd snap trim sleep: 2 m...
705
25.148148
49
yaml
null
ceph-main/qa/suites/rados/thrash/thrashers/default.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost conf: osd: osd debug reject backfill probability: .3 osd scrub min interval: 60 osd scrub max interval: 120 osd max backfills: 3 osd snap trim sleep: 2 ...
763
25.344828
49
yaml
null
ceph-main/qa/suites/rados/thrash/thrashers/mapgap.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost - osd_map_cache_size conf: mon: mon min osdmap epochs: 50 paxos service trim min: 10 # prune full osdmaps regularly mon osdmap full prune min: 15 mon osdma...
729
25.071429
41
yaml
null
ceph-main/qa/suites/rados/thrash/thrashers/morepggrow.yaml
overrides: ceph: conf: osd: osd scrub min interval: 60 osd scrub max interval: 120 journal throttle high multiple: 2 journal throttle max multiple: 10 filestore queue throttle high multiple: 2 filestore queue throttle max multiple: 10 osd max backfills...
540
22.521739
49
yaml
null
ceph-main/qa/suites/rados/thrash/thrashers/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/thrash/thrashers/pggrow.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost conf: osd: osd scrub min interval: 60 osd scrub max interval: 120 filestore odsync write: true osd max backfills: 2 osd snap trim sleep: .5 mon: ...
629
24.2
41
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml
overrides: ceph: conf: client.0: admin socket: /var/run/ceph/ceph-$name.asok tasks: - radosbench: clients: [client.0] time: 150 - admin_socket: client.0: objecter_requests: test: "http://git.ceph.com/?p={repo};a=blob_plain;f=src/test/admin_socket/objecter_requests;hb={branc...
324
22.214286
112
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache-agent-big.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd erasur...
1,167
30.567568
89
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache-agent-small.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd pool c...
1,107
30.657143
78
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd pool c...
1,126
27.175
78
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd pool c...
1,309
28.111111
78
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache-snaps-balanced.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd pool c...
1,197
28.219512
78
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache-snaps.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd pool c...
1,173
28.35
78
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/cache.yaml
overrides: ceph: log-ignorelist: - must scrub before tier agent can activate conf: osd: # override short_pg_log_entries.yaml (which sets these under [global]) osd_min_pg_log_entries: 3000 osd_max_pg_log_entries: 3000 tasks: - exec: client.0: - sudo ceph osd pool c...
1,119
29.27027
78
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/dedup-io-mixed.yaml
tasks: - exec: client.0: - sudo ceph osd pool create low_tier 4 - rados: clients: [client.0] low_tier_pool: 'low_tier' ops: 1500 objects: 50 set_chunk: true enable_dedup: true dedup_chunk_size: '131072' dedup_chunk_algo: 'fastcdc' op_weights: read: 100 write: 50...
405
18.333333
44
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/dedup-io-snaps.yaml
tasks: - exec: client.0: - sudo ceph osd pool create low_tier 4 - rados: clients: [client.0] low_tier_pool: 'low_tier' ops: 1500 objects: 50 set_chunk: true enable_dedup: true dedup_chunk_size: '131072' dedup_chunk_algo: 'fastcdc' op_weights: read: 100 write: 50...
468
18.541667
44
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml
overrides: conf: osd: osd deep scrub update digest min age: 0 tasks: - rados: clients: [client.0] ops: 4000 objects: 50 pool_snaps: true op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 rollback: 50 copy_from: 50
314
15.578947
45
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/rados_api_tests.yaml
overrides: ceph: log-ignorelist: - reached quota - \(POOL_APP_NOT_ENABLED\) - \(PG_AVAILABILITY\) crush_tunables: jewel conf: client: debug ms: 1 debug objecter: 20 debug rados: 20 mon: mon warn on pool no app: false debug mgrc: 20 ...
468
18.541667
38
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/radosbench-high-concurrency.yaml
overrides: ceph: conf: client.0: debug ms: 1 debug objecter: 20 debug rados: 20 tasks: - full_sequential: - radosbench: clients: [client.0] concurrency: 128 size: 8192 time: 90 - radosbench: clients: [client.0] concurrency: 128 size: 8192...
918
17.38
26
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/radosbench.yaml
overrides: ceph: conf: client.0: debug ms: 1 debug objecter: 20 debug rados: 20 tasks: - full_sequential: - radosbench: clients: [client.0] time: 90 - radosbench: clients: [client.0] time: 90 - radosbench: clients: [client.0] time: 90 - rad...
427
16.12
26
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/redirect.yaml
tasks: - exec: client.0: - sudo ceph osd pool create low_tier 4 - rados: clients: [client.0] low_tier_pool: 'low_tier' ops: 4000 objects: 500 set_redirect: true op_weights: read: 100 write: 100 delete: 50 copy_from: 50
277
16.375
44
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/redirect_promote_tests.yaml
tasks: - exec: client.0: - sudo ceph osd pool create low_tier 4 - rados: clients: [client.0] low_tier_pool: 'low_tier' ops: 4000 objects: 500 set_redirect: true op_weights: set_redirect: 100 read: 50 tier_promote: 30
269
17
44
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/redirect_set_object.yaml
tasks: - exec: client.0: - sudo ceph osd pool create low_tier 4 - rados: clients: [client.0] low_tier_pool: 'low_tier' ops: 4000 objects: 500 set_redirect: true op_weights: set_redirect: 100 copy_from: 100
252
17.071429
44
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/set-chunks-read.yaml
tasks: - exec: client.0: - sudo ceph osd pool create low_tier 4 - rados: clients: [client.0] low_tier_pool: 'low_tier' ops: 4000 objects: 300 set_chunk: true op_weights: chunk_read: 100 tier_promote: 10
249
16.857143
44
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/small-objects-balanced.yaml
overrides: ceph: crush_tunables: jewel tasks: - rados: clients: [client.0] ops: 400000 max_seconds: 600 max_in_flight: 64 objects: 1024 size: 16384 balance_reads: true op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 ro...
386
15.826087
25
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/small-objects-localized.yaml
overrides: ceph: crush_tunables: jewel tasks: - rados: clients: [client.0] ops: 400000 max_seconds: 600 max_in_flight: 64 objects: 1024 size: 16384 localize_reads: true op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 r...
387
15.869565
25
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/small-objects.yaml
overrides: ceph: crush_tunables: jewel tasks: - rados: clients: [client.0] ops: 400000 max_seconds: 600 max_in_flight: 64 objects: 1024 size: 16384 op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 rollback: 50 copy_fr...
362
15.5
25
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/snaps-few-objects-balanced.yaml
tasks: - rados: clients: [client.0] ops: 4000 objects: 50 balance_reads: true op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 rollback: 50 copy_from: 50
243
15.266667
23
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/snaps-few-objects-localized.yaml
tasks: - rados: clients: [client.0] ops: 4000 objects: 50 localize_reads: true op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 rollback: 50 copy_from: 50
244
15.333333
24
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml
tasks: - rados: clients: [client.0] ops: 4000 objects: 50 op_weights: read: 100 write: 100 delete: 50 snap_create: 50 snap_remove: 50 rollback: 50 copy_from: 50
219
14.714286
23
yaml
null
ceph-main/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml
tasks: - rados: clients: [client.0] ops: 4000 objects: 500 write_fadvise_dontneed: true op_weights: write: 100
137
14.333333
32
yaml
null
ceph-main/qa/suites/rados/valgrind-leaks/1-start.yaml
openstack: - volumes: # attached to each instance count: 2 size: 10 # GB overrides: install: ceph: debuginfo: true ceph: log-ignorelist: - overall HEALTH_ - \(PG_ conf: global: osd heartbeat grace: 40 osd max object name len: 460 osd max obj...
652
20.064516
69
yaml
null
ceph-main/qa/suites/rados/valgrind-leaks/centos_latest.yaml
.qa/distros/supported/centos_latest.yaml
40
40
40
yaml
null
ceph-main/qa/suites/rados/valgrind-leaks/2-inject-leak/mon.yaml
overrides: ceph: expect_valgrind_errors: true tasks: - exec: mon.a: - ceph tell mon.a leak_some_memory
119
14
40
yaml
null
ceph-main/qa/suites/rados/valgrind-leaks/2-inject-leak/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/valgrind-leaks/2-inject-leak/osd.yaml
overrides: ceph: expect_valgrind_errors: true tasks: - exec: mon.a: - ceph tell osd.0 leak_some_memory
119
14
40
yaml
null
ceph-main/qa/suites/rados/verify/centos_latest.yaml
.qa/distros/supported/centos_latest.yaml
40
40
40
yaml
null
ceph-main/qa/suites/rados/verify/ceph.yaml
overrides: ceph: conf: mon: mon min osdmap epochs: 50 paxos service trim min: 10 # prune full osdmaps regularly mon osdmap full prune min: 15 mon osdmap full prune interval: 2 mon osdmap full prune txsize: 2 osd: debug monc: 20 tasks: - install: ...
328
19.5625
41
yaml
null
ceph-main/qa/suites/rados/verify/rados.yaml
.qa/config/rados.yaml
21
21
21
yaml
null
ceph-main/qa/suites/rados/verify/clusters/fixed-2.yaml
.qa/clusters/fixed-2.yaml
25
25
25
yaml
null
ceph-main/qa/suites/rados/verify/clusters/openstack.yaml
openstack: - volumes: # attached to each instance count: 4 size: 10 # GB
87
16.6
40
yaml
null
ceph-main/qa/suites/rados/verify/d-thrash/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rados/verify/d-thrash/default/default.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost tasks: - thrashosds: timeout: 1200 chance_pgnum_grow: 1 chance_pgnum_shrink: 1 chance_pgpnum_fix: 1
227
18
41
yaml
null
ceph-main/qa/suites/rados/verify/d-thrash/default/thrashosds-health.yaml
.qa/tasks/thrashosds-health.yaml
32
32
32
yaml
null
ceph-main/qa/suites/rados/verify/msgr-failures/few.yaml
overrides: ceph: conf: global: ms inject socket failures: 5000 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
177
18.777778
44
yaml
null
ceph-main/qa/suites/rados/verify/tasks/mon_recovery.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(MON_DOWN\) - \(OSDMAP_FLAGS\) - \(SMALLER_PGP_NUM\) - \(POOL_APP_NOT_ENABLED\) - \(SLOW OPS\) - slow request tasks: - mon_recovery:
235
17.153846
32
yaml
null
ceph-main/qa/suites/rados/verify/tasks/rados_api_tests.yaml
overrides: ceph: log-ignorelist: - reached quota - overall HEALTH_ - \(CACHE_POOL_NO_HIT_SET\) - \(POOL_FULL\) - \(SMALLER_PGP_NUM\) - \(SLOW_OPS\) - \(CACHE_POOL_NEAR_FULL\) - \(POOL_APP_NOT_ENABLED\) - \(PG_AVAILABILITY\) - \(OBJECT_MISPLACED\) -...
743
20.257143
38
yaml
null
ceph-main/qa/suites/rados/verify/tasks/rados_cls_all.yaml
overrides: ceph: conf: osd: osd_class_load_list: "*" osd_class_default_list: "*" tasks: - workunit: clients: client.0: - cls
171
13.333333
35
yaml
null
ceph-main/qa/suites/rados/verify/validater/lockdep.yaml
overrides: ceph: conf: global: lockdep: true
65
10
21
yaml
null
ceph-main/qa/suites/rados/verify/validater/valgrind.yaml
# see http://tracker.ceph.com/issues/20360 and http://tracker.ceph.com/issues/18126 os_type: centos overrides: install: ceph: debuginfo: true ceph: conf: global: osd heartbeat grace: 80 mon: mon osd crush smoke test: false osd: osd fast shutdown: false ...
905
27.3125
150
yaml
null
ceph-main/qa/suites/rbd/basic/base/install.yaml
tasks: - install: - ceph:
26
5.75
10
yaml
null
ceph-main/qa/suites/rbd/basic/cachepool/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rbd/basic/cachepool/small.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(CACHE_POOL_NEAR_FULL\) - \(CACHE_POOL_NO_HIT_SET\) tasks: - exec: client.0: - sudo ceph osd pool create cache 4 - sudo ceph osd tier add rbd cache - sudo ceph osd tier cache-mode cache writeback - sudo ceph osd ti...
569
30.666667
59
yaml
null
ceph-main/qa/suites/rbd/basic/clusters/fixed-1.yaml
.qa/clusters/fixed-1.yaml
25
25
25
yaml
null
ceph-main/qa/suites/rbd/basic/clusters/openstack.yaml
openstack: - volumes: # attached to each instance count: 3 size: 30 # GB
87
16.6
40
yaml
null
ceph-main/qa/suites/rbd/basic/msgr-failures/few.yaml
overrides: ceph: conf: global: ms inject socket failures: 5000 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
177
18.777778
44
yaml
null
ceph-main/qa/suites/rbd/basic/tasks/rbd_api_tests_old_format.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(CACHE_POOL_NO_HIT_SET\) - \(POOL_APP_NOT_ENABLED\) - is full \(reached quota - \(POOL_FULL\) tasks: - workunit: clients: client.0: - rbd/test_librbd.sh
261
17.714286
33
yaml
null
ceph-main/qa/suites/rbd/basic/tasks/rbd_cls_tests.yaml
tasks: - workunit: clients: client.0: - cls/test_cls_rbd.sh - cls/test_cls_lock.sh - cls/test_cls_journal.sh
143
17
33
yaml
null
ceph-main/qa/suites/rbd/basic/tasks/rbd_lock_and_fence.yaml
tasks: - workunit: clients: client.0: - rbd/test_lock_fence.sh
81
12.666667
32
yaml
null
ceph-main/qa/suites/rbd/basic/tasks/rbd_python_api_tests_old_format.yaml
overrides: ceph: log-ignorelist: - \(SLOW_OPS\) - slow request tasks: - workunit: clients: client.0: - rbd/test_librbd_python.sh
165
14.090909
35
yaml
null
ceph-main/qa/suites/rbd/cli/base/install.yaml
tasks: - install: - ceph:
26
5.75
10
yaml
null
ceph-main/qa/suites/rbd/cli/features/defaults.yaml
overrides: ceph: conf: client: rbd default features: 61
76
11.833333
32
yaml
null
ceph-main/qa/suites/rbd/cli/features/journaling.yaml
overrides: ceph: conf: client: rbd default features: 125
77
12
33
yaml
null
ceph-main/qa/suites/rbd/cli/features/layering.yaml
overrides: ceph: conf: client: rbd default features: 1
75
11.666667
31
yaml
null
ceph-main/qa/suites/rbd/cli/msgr-failures/few.yaml
overrides: ceph: conf: global: ms inject socket failures: 5000 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
177
18.777778
44
yaml
null
ceph-main/qa/suites/rbd/cli/pool/ec-data-pool.yaml
tasks: - exec: client.0: - sudo ceph osd erasure-code-profile set teuthologyprofile crush-failure-domain=osd m=1 k=2 - sudo ceph osd pool create datapool 4 4 erasure teuthologyprofile - sudo ceph osd pool set datapool allow_ec_overwrites true - rbd pool init datapool overrides: thrashosds...
951
33
97
yaml
null
ceph-main/qa/suites/rbd/cli/pool/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rbd/cli/pool/replicated-data-pool.yaml
tasks: - exec: client.0: - sudo ceph osd pool create datapool 4 - rbd pool init datapool overrides: ceph: conf: client: rbd default data pool: datapool
189
14.833333
44
yaml
null
ceph-main/qa/suites/rbd/cli/pool/small-cache-pool.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(CACHE_POOL_NEAR_FULL\) - \(CACHE_POOL_NO_HIT_SET\) tasks: - exec: client.0: - sudo ceph osd pool create cache 4 - sudo ceph osd tier add rbd cache - sudo ceph osd tier cache-mode cache writeback - sudo ceph osd ti...
569
30.666667
59
yaml
null
ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_generic.yaml
tasks: - workunit: clients: client.0: - rbd/cli_generic.sh
77
12
28
yaml
null
ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_groups.yaml
tasks: - workunit: clients: client.0: - rbd/rbd_groups.sh
76
11.833333
27
yaml
null
ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_import_export.yaml
tasks: - workunit: clients: client.0: - rbd/import_export.sh
79
12.333333
30
yaml
null
ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_luks_encryption.yaml
overrides: install: ceph: extra_packages: [rbd-nbd] tasks: - workunit: clients: client.0: - rbd/luks-encryption.sh
145
13.6
32
yaml
null
ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_migration.yaml
tasks: - workunit: clients: client.0: - rbd/cli_migration.sh
79
12.333333
30
yaml
null
ceph-main/qa/suites/rbd/cli_v1/base/install.yaml
tasks: - install: - ceph:
26
5.75
10
yaml
null
ceph-main/qa/suites/rbd/cli_v1/features/format-1.yaml
overrides: ceph: conf: client: rbd default format: 1
73
11.333333
29
yaml
null
ceph-main/qa/suites/rbd/cli_v1/msgr-failures/few.yaml
overrides: ceph: conf: global: ms inject socket failures: 5000 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
177
18.777778
44
yaml
null
ceph-main/qa/suites/rbd/cli_v1/pool/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/rbd/cli_v1/pool/small-cache-pool.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(CACHE_POOL_NEAR_FULL\) - \(CACHE_POOL_NO_HIT_SET\) tasks: - exec: client.0: - sudo ceph osd pool create cache 4 - sudo ceph osd tier add rbd cache - sudo ceph osd tier cache-mode cache writeback - sudo ceph osd ti...
569
30.666667
59
yaml