Search is not available for this dataset
repo
stringlengths
2
152
file
stringlengths
15
239
code
stringlengths
0
58.4M
file_length
int64
0
58.4M
avg_line_length
float64
0
1.81M
max_line_length
int64
0
12.7M
extension_type
stringclasses
364 values
null
ceph-main/qa/suites/fs/mixed-clients/distro/centos_8.yaml
.qa/distros/all/centos_8.yaml
29
29
29
yaml
null
ceph-main/qa/suites/fs/mixed-clients/distro/rhel_8.yaml
.qa/distros/all/rhel_8.yaml
27
27
27
yaml
null
ceph-main/qa/suites/fs/mixed-clients/distro/ubuntu/latest.yaml
.qa/distros/all/ubuntu_22.04.yaml
33
33
33
yaml
null
ceph-main/qa/suites/fs/mixed-clients/distro/ubuntu/overrides.yaml
overrides: ceph: valgrind: exit_on_first_error: false
66
12.4
32
yaml
null
ceph-main/qa/suites/fs/mixed-clients/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/mixed-clients/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/mixed-clients/overrides/osd-asserts.yaml
.qa/cephfs/overrides/osd-asserts.yaml
37
37
37
yaml
null
ceph-main/qa/suites/fs/mixed-clients/tasks/kernel_cfuse_workunits_dbench_iozone.yaml
tasks: - parallel: - user-workload - kclient-workload user-workload: sequential: - ceph-fuse: [client.0] - workunit: clients: client.0: - suites/iozone.sh kclient-workload: sequential: - kclient: [client.1] - workunit: clients: client.1: - suites/dbench.sh
327
16.263158
29
yaml
null
ceph-main/qa/suites/fs/mixed-clients/tasks/kernel_cfuse_workunits_untarbuild_blogbench.yaml
tasks: - parallel: - user-workload - kclient-workload user-workload: sequential: - ceph-fuse: [client.0] - workunit: clients: client.0: - suites/blogbench.sh kclient-workload: sequential: - kclient: [client.1] - workunit: clients: client.1: - kernel_untar_build.sh
335
16.684211
34
yaml
null
ceph-main/qa/suites/fs/multiclient/clusters/1-mds-2-client.yaml
.qa/cephfs/clusters/1-mds-2-client.yaml
39
39
39
yaml
null
ceph-main/qa/suites/fs/multiclient/clusters/1-mds-3-client.yaml
.qa/cephfs/clusters/1-mds-3-client.yaml
39
39
39
yaml
null
ceph-main/qa/suites/fs/multiclient/distros/ubuntu_latest.yaml
.qa/distros/supported/ubuntu_latest.yaml
40
40
40
yaml
null
ceph-main/qa/suites/fs/multiclient/mount/fuse.yaml
.qa/cephfs/mount/fuse.yaml
26
26
26
yaml
null
ceph-main/qa/suites/fs/multiclient/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/multiclient/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/multiclient/tasks/cephfs_misc_tests.yaml
tasks: - cephfs_test_runner: modules: - tasks.cephfs.test_misc overrides: ceph: log-ignorelist: - evicting unresponsive client - POOL_APP_NOT_ENABLED - has not responded to cap revoke by MDS for over - MDS_CLIENT_LATE_RELEASE - responding to mclientcaps - RECENT_CRASH
322
20.533333
55
yaml
null
ceph-main/qa/suites/fs/multiclient/tasks/ior-shared-file.yaml
# make sure we get the same MPI version on all hosts tasks: - pexec: clients: - cd $TESTDIR - wget http://download.ceph.com/qa/ior.tbz2 - tar xvfj ior.tbz2 - cd ior - ./configure - make - make install DESTDIR=$TESTDIR/binary/ - cd $TESTDIR/ - rm ior.tbz2 - rm -r ior - ln -s $TESTDIR/mnt.* $TESTDIR/gmnt - ssh_keys: - mpi: exec: $TESTDIR/binary/usr/local/bin/ior -e -w -r -W -b 10m -a POSIX -o $TESTDIR/gmnt/ior.testfile - pexec: clients: - rm -f $TESTDIR/gmnt/ior.testfile - rm -f $TESTDIR/gmnt - rm -rf $TESTDIR/binary
617
24.75
101
yaml
null
ceph-main/qa/suites/fs/multiclient/tasks/mdtest.yaml
# make sure we get the same MPI version on all hosts tasks: - pexec: clients: - cd $TESTDIR - wget http://download.ceph.com/qa/mdtest-1.9.3.tgz - mkdir mdtest-1.9.3 - cd mdtest-1.9.3 - tar xvfz $TESTDIR/mdtest-1.9.3.tgz - rm $TESTDIR/mdtest-1.9.3.tgz - MPI_CC=mpicc make - ln -s $TESTDIR/mnt.* $TESTDIR/gmnt - ssh_keys: - mpi: exec: $TESTDIR/mdtest-1.9.3/mdtest -d $TESTDIR/gmnt -I 20 -z 5 -b 2 -R - pexec: clients: - rm -f $TESTDIR/gmnt - rm -rf $TESTDIR/mdtest-1.9.3 - rm -rf $TESTDIR/._mdtest-1.9.3
580
26.666667
74
yaml
null
ceph-main/qa/suites/fs/multifs/clusters/1a3s-mds-2c-client.yaml
.qa/cephfs/clusters/1a3s-mds-2c-client.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/multifs/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/multifs/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/multifs/overrides/mon-debug.yaml
overrides: ceph: conf: mon: debug mon: 20
62
9.5
21
yaml
null
ceph-main/qa/suites/fs/multifs/tasks/failover.yaml
overrides: ceph: log-ignorelist: - Replacing daemon mds - \(MDS_INSUFFICIENT_STANDBY\) - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - \(MDS_DAMAGE\) - \(FS_DEGRADED\) ceph-fuse: disabled: true tasks: - exec: mon.a: - ceph config set mgr mgr/crash/warn_recent_interval 0 - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_failover
438
19.904762
62
yaml
null
ceph-main/qa/suites/fs/multifs/tasks/multifs-auth.yaml
tasks: - cephfs_test_runner: fail_on_skip: false modules: - tasks.cephfs.test_multifs_auth
113
18
40
yaml
null
ceph-main/qa/suites/fs/permission/clusters/fixed-2-ucephfs.yaml
.qa/cephfs/clusters/fixed-2-ucephfs.yaml
40
40
40
yaml
null
ceph-main/qa/suites/fs/permission/mount/fuse.yaml
.qa/cephfs/mount/fuse.yaml
26
26
26
yaml
null
ceph-main/qa/suites/fs/permission/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/permission/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/permission/tasks/cfuse_workunit_misc.yaml
overrides: ceph: conf: client: client acl type: posix_acl tasks: - workunit: clients: all: - fs/misc/acl.sh - fs/misc/chmod.sh - fs/misc/dac_override.sh
207
15
34
yaml
null
ceph-main/qa/suites/fs/permission/tasks/cfuse_workunit_suites_pjd.yaml
overrides: ceph: conf: client: fuse set user groups: true client acl type: posix_acl tasks: - workunit: timeout: 6h clients: all: - suites/pjd.sh
196
14.153846
34
yaml
null
ceph-main/qa/suites/fs/shell/clusters/1-mds-1-client-coloc.yaml
.qa/cephfs/clusters/1-mds-1-client-coloc.yaml
45
45
45
yaml
null
ceph-main/qa/suites/fs/shell/mount/fuse.yaml
.qa/cephfs/mount/fuse.yaml
26
26
26
yaml
null
ceph-main/qa/suites/fs/shell/objectstore/bluestore-bitmap.yaml
.qa/objectstore/bluestore-bitmap.yaml
37
37
37
yaml
null
ceph-main/qa/suites/fs/shell/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/shell/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/shell/overrides/no_client_pidfile.yaml
.qa/overrides/no_client_pidfile.yaml
36
36
36
yaml
null
ceph-main/qa/suites/fs/shell/tasks/cephfs-shell.yaml
# Right now, cephfs-shell is only available as a package on Ubuntu # This overrides the random distribution that's chosen in the other yaml fragments. os_type: ubuntu os_version: "20.04" tasks: - cephfs_test_runner: modules: - tasks.cephfs.test_cephfs_shell
274
29.555556
83
yaml
null
ceph-main/qa/suites/fs/snaps/clusters/1a3s-mds-1c-client.yaml
.qa/cephfs/clusters/1a3s-mds-1c-client.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/snaps/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/snaps/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/snaps/tasks/workunit/snaps.yaml
tasks: - workunit: timeout: 6h clients: all: - fs/snaps
78
10.285714
18
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/clusters/1a3s-mds-2c-client.yaml
.qa/cephfs/clusters/1a3s-mds-2c-client.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/msgr-failures/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/msgr-failures/osd-mds-delay.yaml
overrides: ceph: conf: osd: ms inject socket failures: 2500 ms inject delay type: client mds ms inject delay probability: .005 ms inject delay max: 1 mon client directed command retry: 5 mds: ms inject socket failures: 2500 ms inject delay type: client mds osd ms inject delay probability: .005 ms inject delay max: 1 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
502
26.944444
44
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/objectstore/bluestore-bitmap.yaml
.qa/objectstore/bluestore-bitmap.yaml
37
37
37
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/overrides/frag.yaml
.qa/cephfs/overrides/frag.yaml
30
30
30
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/overrides/multifs.yaml
overrides: ceph: cephfs: fs: - name: a - name: b ceph-fuse: client.0: cephfs_name: a client.1: cephfs_name: b kclient: client.0: cephfs_name: a client.1: cephfs_name: b
241
13.235294
20
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/overrides/session_timeout.yaml
.qa/cephfs/overrides/session_timeout.yaml
41
41
41
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/overrides/thrashosds-health.yaml
.qa/tasks/thrashosds-health.yaml
32
32
32
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/1-thrash/mds.yaml
tasks: - mds_thrash: overrides: ceph: log-ignorelist: - Replacing daemon mds
90
10.375
28
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/1-thrash/mon.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(MON_DOWN\) tasks: - mon_thrash: check_mds_failover: True revive_delay: 20 thrash_delay: 10
176
15.090909
28
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/2-workunit/cfuse_workunit_snaptests.yaml
tasks: - workunit: timeout: 6h clients: all: - fs/snaps
78
10.285714
18
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/2-workunit/cfuse_workunit_suites_fsstress.yaml
.qa/cephfs/tasks/cfuse_workunit_suites_fsstress.yaml
52
52
52
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/2-workunit/cfuse_workunit_suites_pjd.yaml
overrides: ceph: conf: client: fuse set user groups: true tasks: - workunit: timeout: 6h clients: all: - suites/pjd.sh
161
12.5
34
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/2-workunit/cfuse_workunit_trivial_sync.yaml
.qa/cephfs/tasks/cfuse_workunit_trivial_sync.yaml
49
49
49
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/2-workunit/ffsb.yaml
overrides: ceph: log-ignorelist: - SLOW_OPS - slow request conf: osd: filestore flush min: 0 tasks: - workunit: clients: all: - suites/ffsb.sh
193
12.857143
30
yaml
null
ceph-main/qa/suites/fs/thrash/multifs/tasks/2-workunit/iozone.yaml
tasks: - workunit: clients: all: - suites/iozone.sh
70
10.833333
26
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/clusters/1a5s-mds-1c-client.yaml
.qa/cephfs/clusters/1a5s-mds-1c-client.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/msgr-failures/none.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/msgr-failures/osd-mds-delay.yaml
overrides: ceph: conf: osd: ms inject socket failures: 2500 ms inject delay type: client mds ms inject delay probability: .005 ms inject delay max: 1 mon client directed command retry: 5 mds: ms inject socket failures: 2500 ms inject delay type: client mds osd ms inject delay probability: .005 ms inject delay max: 1 mon client directed command retry: 5 log-ignorelist: - \(OSD_SLOW_PING_TIME
502
26.944444
44
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/frag.yaml
.qa/cephfs/overrides/frag.yaml
30
30
30
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/races.yaml
overrides: ceph: conf: mds: mds_sleep_rank_change: 5000000.0
81
12.666667
40
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/session_timeout.yaml
.qa/cephfs/overrides/session_timeout.yaml
41
41
41
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/thrashosds-health.yaml
.qa/tasks/thrashosds-health.yaml
32
32
32
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/prefetch_dirfrags/no.yaml
overrides: ceph: conf: mds: mds oft prefetch dirfrags: false
81
12.666667
40
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/overrides/prefetch_dirfrags/yes.yaml
overrides: ceph: conf: mds: mds oft prefetch dirfrags: true
80
12.5
39
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/ranks/1.yaml
0
0
0
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/ranks/3.yaml
overrides: ceph: cephfs: max_mds: 3 check-counter: counters: mds: - mds.exported - mds.imported
136
12.7
22
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/ranks/5.yaml
overrides: ceph: cephfs: max_mds: 5 check-counter: counters: mds: - mds.exported - mds.imported
136
12.7
22
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/1-thrash/mds.yaml
tasks: - mds_thrash: overrides: ceph: log-ignorelist: - Replacing daemon mds
90
10.375
28
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/1-thrash/mon.yaml
overrides: ceph: log-ignorelist: - overall HEALTH_ - \(MON_DOWN\) tasks: - mon_thrash: check_mds_failover: True revive_delay: 20 thrash_delay: 10
176
15.090909
28
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/1-thrash/osd.yaml
overrides: ceph: log-ignorelist: - but it is still running - objects unfound and apparently lost - MDS_SLOW_METADATA_IO - MDS_TRIM tasks: - thrashosds:
182
17.3
43
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/2-workunit/fs/snaps.yaml
tasks: - workunit: timeout: 6h clients: all: - fs/snaps
78
10.285714
18
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/2-workunit/fs/trivial_sync.yaml
.qa/cephfs/tasks/cfuse_workunit_trivial_sync.yaml
49
49
49
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/2-workunit/suites/ffsb.yaml
overrides: ceph: log-ignorelist: - SLOW_OPS - slow request conf: osd: filestore flush min: 0 tasks: - workunit: clients: all: - suites/ffsb.sh
193
12.857143
30
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/2-workunit/suites/fsstress.yaml
tasks: - workunit: timeout: 6h clients: all: - suites/fsstress.sh
88
11.714286
28
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/2-workunit/suites/iozone.yaml
tasks: - workunit: clients: all: - suites/iozone.sh
70
10.833333
26
yaml
null
ceph-main/qa/suites/fs/thrash/workloads/tasks/2-workunit/suites/pjd.yaml
overrides: ceph: conf: client: fuse set user groups: true tasks: - workunit: timeout: 6h clients: all: - suites/pjd.sh
161
12.5
34
yaml
null
ceph-main/qa/suites/fs/top/cluster/1-node.yaml
meta: - desc: 1 ceph cluster with 1 mon, 1 mgr, 3 osds, 2 mds, 2 clients roles: - - mon.a - mgr.x - mds.a - mds.b - osd.0 - osd.1 - osd.2 - client.0 - client.1
176
12.615385
66
yaml
null
ceph-main/qa/suites/fs/top/mount/fuse.yaml
.qa/cephfs/mount/fuse.yaml
26
26
26
yaml
null
ceph-main/qa/suites/fs/top/objectstore/bluestore-bitmap.yaml
.qa/objectstore/bluestore-bitmap.yaml
37
37
37
yaml
null
ceph-main/qa/suites/fs/top/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/top/tasks/fstop.yaml
tasks: - cephfs_test_runner: modules: - tasks.cephfs.test_fstop
80
15.2
33
yaml
null
ceph-main/qa/suites/fs/traceless/clusters/fixed-2-ucephfs.yaml
.qa/cephfs/clusters/fixed-2-ucephfs.yaml
40
40
40
yaml
null
ceph-main/qa/suites/fs/traceless/mount/fuse.yaml
.qa/cephfs/mount/fuse.yaml
26
26
26
yaml
null
ceph-main/qa/suites/fs/traceless/overrides/frag.yaml
.qa/cephfs/overrides/frag.yaml
30
30
30
yaml
null
ceph-main/qa/suites/fs/traceless/overrides/ignorelist_health.yaml
.qa/cephfs/overrides/ignorelist_health.yaml
43
43
43
yaml
null
ceph-main/qa/suites/fs/traceless/overrides/ignorelist_wrongly_marked_down.yaml
.qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml
56
56
56
yaml
null
ceph-main/qa/suites/fs/traceless/tasks/cfuse_workunit_suites_blogbench.yaml
.qa/cephfs/tasks/cfuse_workunit_suites_blogbench.yaml
53
53
53
yaml
null
ceph-main/qa/suites/fs/traceless/tasks/cfuse_workunit_suites_dbench.yaml
.qa/cephfs/tasks/cfuse_workunit_suites_dbench.yaml
50
50
50
yaml
null
ceph-main/qa/suites/fs/traceless/tasks/cfuse_workunit_suites_ffsb.yaml
.qa/cephfs/tasks/cfuse_workunit_suites_ffsb.yaml
48
48
48
yaml
null
ceph-main/qa/suites/fs/traceless/tasks/cfuse_workunit_suites_fsstress.yaml
.qa/cephfs/tasks/cfuse_workunit_suites_fsstress.yaml
52
52
52
yaml
null
ceph-main/qa/suites/fs/traceless/traceless/50pc.yaml
overrides: ceph: conf: mds: mds inject traceless reply probability: .5
91
14.333333
50
yaml
null
ceph-main/qa/suites/fs/upgrade/featureful_client/old_client/bluestore-bitmap.yaml
../../../../../cephfs/objectstore-ec/bluestore-bitmap.yaml
58
58
58
yaml
null
ceph-main/qa/suites/fs/upgrade/featureful_client/old_client/centos_latest.yaml
.qa/distros/supported/centos_latest.yaml
40
40
40
yaml
null
ceph-main/qa/suites/fs/upgrade/featureful_client/old_client/clusters/1-mds-2-client-micro.yaml
.qa/cephfs/clusters/1-mds-2-client-micro.yaml
45
45
45
yaml