Search is not available for this dataset
repo
stringlengths
2
152
file
stringlengths
15
239
code
stringlengths
0
58.4M
file_length
int64
0
58.4M
avg_line_length
float64
0
1.81M
max_line_length
int64
0
12.7M
extension_type
stringclasses
364 values
null
ceph-main/qa/suites/fs/verify/validater/valgrind.yaml
overrides: install: ceph: debuginfo: true ceph: # Valgrind makes everything slow, so ignore slow requests and extend heartbeat grace log-ignorelist: - slow request - SLOW_OPS - MON_DOWN conf: global: osd heartbeat grace: 60 mds heartbeat grace: 60 ...
739
23.666667
88
yaml
null
ceph-main/qa/suites/fs/volumes/clusters/1a3s-mds-4c-client.yaml
.qa/cephfs/clusters/1a3s-mds-4c-client.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/volumes/objectstore/bluestore-bitmap.yaml
.qa/objectstore/bluestore-bitmap.yaml
37
37
37
yaml
null
ceph-main/qa/suites/fs/volumes/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/volumes/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/volumes/overrides/no_client_pidfile.yaml
.qa/overrides/no_client_pidfile.yaml
36
36
36
yaml
null
ceph-main/qa/suites/fs/volumes/tasks/volumes/overrides.yaml
overrides: ceph: conf: mgr: debug client: 20 debug ms: 1 debug finisher: 20 debug mgr: 20 log-ignorelist: - OSD full dropping all updates - OSD near full - pausewr flag - failsafe engaged, dropping updates - failsafe disengaged, no longer dro...
402
21.388889
47
yaml
null
ceph-main/qa/suites/fs/volumes/tasks/volumes/test/basic.yaml
tasks: - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_volumes.TestVolumes - tasks.cephfs.test_volumes.TestSubvolumeGroups - tasks.cephfs.test_volumes.TestSubvolumes - tasks.cephfs.test_subvolume.TestSubvolume
279
30.111111
55
yaml
null
ceph-main/qa/suites/fs/volumes/tasks/volumes/test/clone.yaml
tasks: - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_volumes.TestSubvolumeSnapshotClones
136
21.833333
63
yaml
null
ceph-main/qa/suites/fs/volumes/tasks/volumes/test/finisher_per_module.yaml
tasks: - check-counter: counters: mgr: - name: "finisher-volumes.complete_latency.avgcount" min: 4 - name: "finisher-volumes.queue_len" expected_val: 0 - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_volume...
349
24
64
yaml
null
ceph-main/qa/suites/fs/volumes/tasks/volumes/test/misc.yaml
tasks: - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_volumes.TestMisc
117
18.666667
44
yaml
null
ceph-main/qa/suites/fs/volumes/tasks/volumes/test/snapshot.yaml
tasks: - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_volumes.TestSubvolumeGroupSnapshots - tasks.cephfs.test_volumes.TestSubvolumeSnapshots
195
27
63
yaml
null
ceph-main/qa/suites/fs/workload/0-rhel_8.yaml
.qa/distros/podman/rhel_8.6_container_tools_rhel8.yaml
54
54
54
yaml
null
ceph-main/qa/suites/fs/workload/standby-replay.yaml
overrides: ceph: cephfs: standby_replay: true
58
10.8
26
yaml
null
ceph-main/qa/suites/fs/workload/begin/0-install.yaml
.qa/cephfs/begin/0-install.yaml
31
31
31
yaml
null
ceph-main/qa/suites/fs/workload/begin/1-cephadm.yaml
overrides: ceph: conf: osd: osd shutdown pgref assert: true tasks: - cephadm: roleless: false - cephadm.shell: mon.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - cephadm.shell: mon.a: - ceph fs dump ...
367
15.727273
39
yaml
null
ceph-main/qa/suites/fs/workload/begin/2-logrotate.yaml
.qa/cephfs/begin/2-logrotate.yaml
33
33
33
yaml
null
ceph-main/qa/suites/fs/workload/clusters/1a11s-mds-1c-client-3node.yaml
.qa/cephfs/clusters/1a11s-mds-1c-client-3node.yaml
50
50
50
yaml
null
ceph-main/qa/suites/fs/workload/mount/fuse.yaml
.qa/cephfs/mount/fuse.yaml
26
26
26
yaml
null
ceph-main/qa/suites/fs/workload/mount/kclient/ms_mode/crc.yaml
overrides: kclient: mntopts: ["ms_mode=crc"]
51
12
28
yaml
null
ceph-main/qa/suites/fs/workload/mount/kclient/ms_mode/legacy.yaml
overrides: kclient: mntopts: ["ms_mode=legacy"]
54
12.75
31
yaml
null
ceph-main/qa/suites/fs/workload/mount/kclient/ms_mode/secure.yaml
overrides: kclient: mntopts: ["ms_mode=secure"]
54
12.75
31
yaml
null
ceph-main/qa/suites/fs/workload/mount/kclient/wsync/no.yaml
overrides: kclient: mntopts: ["nowsync"]
47
11
24
yaml
null
ceph-main/qa/suites/fs/workload/mount/kclient/wsync/yes.yaml
overrides: kclient: mntopts: ["wsync"]
45
10.5
22
yaml
null
ceph-main/qa/suites/fs/workload/omap_limit/10.yaml
overrides: ceph: conf: osd: osd_max_omap_entries_per_request: 10
85
13.333333
44
yaml
null
ceph-main/qa/suites/fs/workload/omap_limit/10000.yaml
overrides: ceph: conf: osd: osd_max_omap_entries_per_request: 10000
88
13.833333
47
yaml
null
ceph-main/qa/suites/fs/workload/overrides/cephsqlite-timeout.yaml
# increase lock renewal timeout: OSD stress from small clusters may cause # spurious timeouts overrides: ceph: conf: mgr: cephsqlite lock renewal timeout: 900000
182
21.875
73
yaml
null
ceph-main/qa/suites/fs/workload/overrides/frag.yaml
.qa/cephfs/overrides/frag.yaml
30
30
30
yaml
null
ceph-main/qa/suites/fs/workload/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/workload/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/workload/overrides/osd-asserts.yaml
.qa/cephfs/overrides/osd-asserts.yaml
37
37
37
yaml
null
ceph-main/qa/suites/fs/workload/overrides/session_timeout.yaml
.qa/cephfs/overrides/session_timeout.yaml
41
41
41
yaml
null
ceph-main/qa/suites/fs/workload/ranks/1.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/workload/ranks/multi/export-check.yaml
overrides: check-counter: counters: mds: - mds.exported - mds.imported
99
13.285714
22
yaml
null
ceph-main/qa/suites/fs/workload/ranks/multi/n/3.yaml
overrides: ceph: cephfs: max_mds: 3
48
8.8
16
yaml
null
ceph-main/qa/suites/fs/workload/ranks/multi/n/5.yaml
overrides: ceph: cephfs: max_mds: 5
48
8.8
16
yaml
null
ceph-main/qa/suites/fs/workload/ranks/multi/replication/always.yaml
# To exercise lock/witness code paths more regularly, try to get all # directories replicated. overrides: ceph: conf: mds: mds_bal_replicate_threshold: 1 # Note: dir_update is only sent by an MDS trying to replicate a dir. # dir_update is always sent for root, so the count should be more than 2 ch...
494
25.052632
72
yaml
null
ceph-main/qa/suites/fs/workload/ranks/multi/replication/default.yaml
# Use default (8000)
21
10
20
yaml
null
ceph-main/qa/suites/fs/workload/tasks/1-check-counter.yaml
tasks: - check-counter: {}
27
8.333333
19
yaml
null
ceph-main/qa/suites/fs/workload/tasks/0-subvolume/no-subvolume.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/workload/tasks/0-subvolume/with-namespace-isolated-and-quota.yaml
overrides: ceph: subvols: create: 2 subvol_options: "--namespace-isolated --size 25000000000" ceph-fuse: client.0: mount_subvol_num: 0 kclient: client.0: mount_subvol_num: 1
216
17.083333
63
yaml
null
ceph-main/qa/suites/fs/workload/tasks/0-subvolume/with-namespace-isolated.yaml
overrides: ceph: subvols: create: 2 subvol_options: "--namespace-isolated" ceph-fuse: client.0: mount_subvol_num: 0 kclient: client.0: mount_subvol_num: 1
197
15.5
44
yaml
null
ceph-main/qa/suites/fs/workload/tasks/0-subvolume/with-no-extra-options.yaml
overrides: ceph: subvols: create: 2 ceph-fuse: client.0: mount_subvol_num: 0 kclient: client.0: mount_subvol_num: 1
152
12.909091
25
yaml
null
ceph-main/qa/suites/fs/workload/tasks/0-subvolume/with-quota.yaml
overrides: ceph: subvols: create: 2 subvol_options: "--size 25000000000" ceph-fuse: client.0: mount_subvol_num: 0 kclient: client.0: mount_subvol_num: 1
195
15.333333
42
yaml
null
ceph-main/qa/suites/fs/workload/tasks/2-scrub/no.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/workload/tasks/2-scrub/yes.yaml
overrides: ceph: log-ignorelist: - slow metadata IO - SLOW_OPS - slow request tasks: - fwd_scrub: scrub_timeout: 900 sleep_between_iterations: 1
177
15.181818
31
yaml
null
ceph-main/qa/suites/fs/workload/tasks/3-snaps/no.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/workload/tasks/3-snaps/yes.yaml
overrides: ceph: conf: mgr: debug cephsqlite: 20 check-counter: counters: mds: - mds.root_rsnaps - mds_server.req_mksnap_latency.avgcount - mds_server.req_rmsnap_latency.avgcount tasks: - exec: mon.a: - ceph mgr module enable snap_schedule - ceph c...
1,246
39.225806
162
yaml
null
ceph-main/qa/suites/fs/workload/tasks/4-flush/no.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/workload/tasks/4-flush/yes.yaml
tasks: - background_exec: mon.a: - while sleep 13; do ceph tell mds.cephfs:0 flush journal; done
107
20.6
69
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/direct_io.yaml
overrides: check-counter: dry_run: true tasks: - workunit: clients: all: - direct_io
110
10.1
19
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/kernel_untar_build.yaml
overrides: check-counter: counters: mds: - "mds.dir_split" tasks: - workunit: clients: all: - kernel_untar_build.sh
154
13.090909
31
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/postgres.yaml
# I would expect setting the context on the postgresql database directories # would correctly trickle down to the files created by the postgresql daemon, # but this does not appear to work. I would still see denials like: # type=AVC msg=audit(1655861665.521:21354): avc: denied { create } for pid=131994 comm="po...
2,056
54.594595
248
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/fs/misc.yaml
overrides: check-counter: counters: mds: - "mds.dir_split" tasks: - workunit: clients: all: - fs/misc
140
11.818182
25
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/fs/norstats.yaml
overrides: check-counter: counters: mds: - "mds.dir_split" tasks: - workunit: clients: all: - fs/norstats overrides: kclient: rbytes: false ceph: conf: client: client dirsize rbytes: false
253
13.111111
36
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/fs/test_o_trunc.yaml
overrides: check-counter: dry_run: true tasks: - workunit: clients: all: - fs/test_o_trunc.sh
118
12.222222
28
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/blogbench.yaml
overrides: check-counter: counters: mds: - "mds.dir_split" tasks: - workunit: clients: all: - suites/blogbench.sh
152
12.909091
29
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/dbench.yaml
tasks: - workunit: clients: all: - suites/dbench.sh
70
10.833333
26
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/ffsb.yaml
overrides: ceph: log-ignorelist: - SLOW_OPS - slow request conf: osd: filestore flush min: 0 osd heartbeat grace: 60 check-counter: counters: mds: - "mds.dir_split" tasks: - workunit: clients: all: - suites/ffsb.sh
293
14.473684
31
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/fsstress.yaml
tasks: - workunit: timeout: 6h clients: all: - suites/fsstress.sh
88
11.714286
28
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/fsx.yaml
overrides: check-counter: counters: mds: - "mds.dir_split" tasks: - workunit: clients: all: - suites/fsx.sh
146
12.363636
25
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/fsync-tester.yaml
overrides: check-counter: dry_run: true tasks: - workunit: clients: all: - suites/fsync-tester.sh
122
12.666667
32
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/iogen.yaml
overrides: check-counter: dry_run: true tasks: - workunit: clients: all: - suites/iogen.sh
115
11.888889
25
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/iozone.yaml
overrides: check-counter: dry_run: true tasks: - workunit: clients: all: - suites/iozone.sh
116
12
26
yaml
null
ceph-main/qa/suites/fs/workload/tasks/5-workunit/suites/pjd.yaml
overrides: ceph: conf: client: fuse set user groups: true fuse default permissions: false tasks: - workunit: timeout: 6h clients: all: - suites/pjd.sh
201
14.538462
39
yaml
null
ceph-main/qa/suites/hadoop/basic/clusters/fixed-3.yaml
overrides: ceph: conf: client: client permissions: false roles: - [mon.0, mds.a, osd.0, hadoop.master.0] - [mon.1, mgr.x, osd.1, hadoop.slave.0] - [mon.2, mgr.y, hadoop.slave.1, client.0] openstack: - volumes: # attached to each instance count: 1 size: 10 # GB
289
19.714286
42
yaml
null
ceph-main/qa/suites/hadoop/basic/distros/ubuntu_latest.yaml
.qa/distros/supported/ubuntu_latest.yaml
40
40
40
yaml
null
ceph-main/qa/suites/hadoop/basic/tasks/repl.yaml
tasks: - ssh_keys: - install: - ceph: - hadoop: - workunit: clients: client.0: [hadoop/repl.sh]
106
10.888889
32
yaml
null
ceph-main/qa/suites/hadoop/basic/tasks/terasort.yaml
tasks: - ssh_keys: - install: - ceph: - hadoop: - workunit: clients: client.0: [hadoop/terasort.sh] env: NUM_RECORDS: "10000000"
150
12.727273
36
yaml
null
ceph-main/qa/suites/hadoop/basic/tasks/wordcount.yaml
tasks: - ssh_keys: - install: - ceph: - hadoop: - workunit: clients: client.0: [hadoop/wordcount.sh]
112
11.555556
37
yaml
null
ceph-main/qa/suites/krbd/basic/bluestore-bitmap.yaml
.qa/objectstore/bluestore-bitmap.yaml
37
37
37
yaml
null
ceph-main/qa/suites/krbd/basic/conf.yaml
overrides: ceph: conf: global: ms die on skipped message: false client: rbd default features: 37
131
15.5
40
yaml
null
ceph-main/qa/suites/krbd/basic/ceph/ceph.yaml
tasks: - install: - ceph:
26
5.75
10
yaml
null
ceph-main/qa/suites/krbd/basic/clusters/fixed-1.yaml
.qa/clusters/fixed-1.yaml
25
25
25
yaml
null
ceph-main/qa/suites/krbd/basic/ms_mode/secure.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=secure
91
14.333333
47
yaml
null
ceph-main/qa/suites/krbd/basic/ms_mode/crc$/crc-rxbounce.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=crc,rxbounce
97
15.333333
53
yaml
null
ceph-main/qa/suites/krbd/basic/ms_mode/crc$/crc.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=crc
88
13.833333
44
yaml
null
ceph-main/qa/suites/krbd/basic/ms_mode/legacy$/legacy-rxbounce.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=legacy,rxbounce
100
15.833333
56
yaml
null
ceph-main/qa/suites/krbd/basic/ms_mode/legacy$/legacy.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=legacy
91
14.333333
47
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_deep_flatten.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_deep_flatten.t
79
12.333333
34
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_discard.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_discard.t - qa/rbd/krbd_discard_512b.t - qa/rbd/krbd_discard_4M.t - qa/rbd/krbd_zeroout.t - qa/rbd/krbd_discard_granularity.t
214
20.5
41
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_huge_image.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_huge_image.t
77
12
32
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_modprobe.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_modprobe.t
75
11.666667
30
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_msgr_segments.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_msgr_segments.t
80
12.5
35
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_parent_overlap.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_parent_overlap.t
81
12.666667
36
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_read_only.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_blkroset.t - qa/rbd/krbd_get_features.t
110
14.857143
34
yaml
null
ceph-main/qa/suites/krbd/basic/tasks/krbd_whole_object_zeroout.yaml
tasks: - cram: clients: client.0: - qa/rbd/krbd_whole_object_zeroout.t
87
13.666667
42
yaml
null
ceph-main/qa/suites/krbd/fsx/conf.yaml
overrides: ceph: conf: global: ms die on skipped message: false
84
13.166667
40
yaml
null
ceph-main/qa/suites/krbd/fsx/ceph/ceph.yaml
tasks: - install: - ceph:
26
5.75
10
yaml
null
ceph-main/qa/suites/krbd/fsx/clusters/3-node.yaml
# fixed-3.yaml, but with two additional clients on the same target roles: - [mon.a, mon.c, mgr.x, osd.0, osd.1, osd.2, osd.3] - [mon.b, mgr.y, osd.4, osd.5, osd.6, osd.7] - [client.0, client.1, client.2] openstack: - volumes: # attached to each instance count: 4 size: 10 # GB overrides: ceph: conf: ...
365
23.4
66
yaml
null
ceph-main/qa/suites/krbd/fsx/features/no-object-map.yaml
overrides: ceph: conf: client: # layering, deep-flatten rbd default features: 33
109
14.714286
32
yaml
null
ceph-main/qa/suites/krbd/fsx/features/object-map.yaml
overrides: ceph: conf: client: # layering, exclusive-lock, object-map, fast-diff, deep-flatten rbd default features: 61
148
20.285714
71
yaml
null
ceph-main/qa/suites/krbd/fsx/ms_mode$/crc-rxbounce.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=crc,rxbounce,read_from_replica=balance
123
19.666667
79
yaml
null
ceph-main/qa/suites/krbd/fsx/ms_mode$/crc.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=crc,read_from_replica=balance
114
18.166667
70
yaml
null
ceph-main/qa/suites/krbd/fsx/ms_mode$/legacy-rxbounce.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=legacy,rxbounce,read_from_replica=balance
126
20.166667
82
yaml
null
ceph-main/qa/suites/krbd/fsx/ms_mode$/legacy.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=legacy,read_from_replica=balance
117
18.666667
73
yaml
null
ceph-main/qa/suites/krbd/fsx/ms_mode$/secure.yaml
overrides: ceph: conf: client: rbd default map options: ms_mode=secure,read_from_replica=balance
117
18.666667
73
yaml
null
ceph-main/qa/suites/krbd/fsx/objectstore/bluestore-bitmap.yaml
.qa/objectstore/bluestore-bitmap.yaml
37
37
37
yaml
null
ceph-main/qa/suites/krbd/fsx/striping/default/randomized-striping-off.yaml
overrides: rbd_fsx: randomized_striping: false
53
12.5
30
yaml
null
ceph-main/qa/suites/krbd/fsx/striping/default/msgr-failures/few.yaml
overrides: ceph: conf: global: ms inject socket failures: 5000 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
177
18.777778
44
yaml