Search is not available for this dataset
repo
stringlengths 2
152
⌀ | file
stringlengths 15
239
| code
stringlengths 0
58.4M
| file_length
int64 0
58.4M
| avg_line_length
float64 0
1.81M
| max_line_length
int64 0
12.7M
| extension_type
stringclasses 364
values |
---|---|---|---|---|---|---|
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/overrides/ignorelist_wrongly_marked_down.yaml | .qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml | 56 | 56 | 56 | yaml |
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/overrides/no_client_pidfile.yaml | .qa/overrides/no_client_pidfile.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/tasks/trim-i24137.yaml | # Note this test is unlikely to exercise the code as expected in the future:
# "It's too tricky to arrange inodes in session->caps. we don't know if it
# still works in the future." -Zheng
tasks:
- exec:
mon.a:
- ceph config set mds mds_min_caps_per_client 1
- background_exec:
mon.a:
- "sleep 30 && ceph config set mds mds_max_caps_per_client 1"
- exec:
client.0:
- ceph_test_trim_caps
| 412 | 26.533333 | 76 | yaml |
null | ceph-main/qa/suites/fs/cephadm/multivolume/0-start.yaml | roles:
- - host.a
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- - host.b
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
openstack:
- volumes: # attached to each instance
count: 4
size: 10 # GB
overrides:
ceph:
conf:
osd:
osd shutdown pgref assert: true
tasks:
- install:
- cephadm:
roleless: true
- cephadm.shell:
host.a:
- ceph orch status
- ceph orch ps
- ceph orch ls
- ceph orch host ls
- ceph orch device ls
- cephadm.shell:
host.a:
- ceph fs volume create foo
- ceph fs volume create bar
- fs.ready:
timeout: 300
| 620 | 14.525 | 39 | yaml |
null | ceph-main/qa/suites/fs/cephadm/multivolume/1-mount.yaml | tasks:
- ceph-fuse:
client.0:
cephfs_name: foo
- ceph-fuse:
client.1:
cephfs_name: bar
| 119 | 14 | 24 | yaml |
null | ceph-main/qa/suites/fs/cephadm/multivolume/2-workload/dbench.yaml | .qa/suites/fs/workload/tasks/5-workunit/suites/dbench.yaml | 58 | 58 | 58 | yaml |
null | ceph-main/qa/suites/fs/cephadm/multivolume/distro/single-container-host.yaml | .qa/distros/single-container-host.yaml | 38 | 38 | 38 | yaml |
null | ceph-main/qa/suites/fs/cephadm/renamevolume/0-start.yaml | roles:
- - host.a
- osd.0
- osd.1
- osd.2
- osd.3
- client.0
- - host.b
- osd.4
- osd.5
- osd.6
- osd.7
- client.1
openstack:
- volumes: # attached to each instance
count: 4
size: 10 # GB
overrides:
ceph:
conf:
osd:
osd shutdown pgref assert: true
tasks:
- install:
- cephadm:
roleless: true
- cephadm.shell:
host.a:
- ceph orch status
- ceph orch ps
- ceph orch ls
- ceph orch host ls
- ceph orch device ls
- cephadm.shell:
host.a:
- ceph fs volume create foo
- fs.ready:
timeout: 300
| 586 | 14.051282 | 39 | yaml |
null | ceph-main/qa/suites/fs/cephadm/renamevolume/1-rename.yaml | tasks:
- cephadm.shell:
host.a:
- ceph fs volume rename foo bar --yes-i-really-mean-it
- fs.ready:
timeout: 300
- cephadm.shell:
host.a:
- |
set -ex
ceph orch ls mds --format=json | jq ".[] | .service_name" | grep "mds.bar"
| 264 | 21.083333 | 82 | yaml |
null | ceph-main/qa/suites/fs/cephadm/renamevolume/distro/single-container-host.yaml | .qa/distros/single-container-host.yaml | 38 | 38 | 38 | yaml |
null | ceph-main/qa/suites/fs/cephadm/renamevolume/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/bluestore-bitmap.yaml | .qa/cephfs/objectstore-ec/bluestore-bitmap.yaml | 47 | 47 | 47 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/clusters/1-mds-1-client.yaml | .qa/cephfs/clusters/1-mds-1-client.yaml | 39 | 39 | 39 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/distro/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/distro/ubuntu_latest.yaml | .qa/distros/supported/ubuntu_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/overrides/ignorelist_health_more.yaml | overrides:
ceph:
log-ignorelist:
- Reduced data availability
- Degraded data redundancy
| 106 | 16.833333 | 33 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/overrides/ignorelist_wrongly_marked_down.yaml | .qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml | 56 | 56 | 56 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/overrides/pg-warn.yaml | overrides:
ceph:
conf:
global:
mon pg warn min per osd: 0
| 78 | 12.166667 | 34 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/tasks/0-client.yaml | teuthology:
postmerge:
# Once can we make sure the distro kernels have included the fscrypt feature
# or the ceph-fuse have supported the fscrypt feature we can remove this
# restriction.
- if not is_kupstream() then reject() end
task:
| 254 | 30.875 | 80 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/tasks/1-tests/fscrypt-common.yaml | tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_fscrypt
| 108 | 17.166667 | 35 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/tasks/1-tests/fscrypt-dbench.yaml | tasks:
- workunit:
timeout: 6h
clients:
client.0:
- fs/fscrypt.sh none dbench
- fs/fscrypt.sh unlocked dbench
| 140 | 16.625 | 39 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/tasks/1-tests/fscrypt-ffsb.yaml | tasks:
- workunit:
timeout: 6h
clients:
client.0:
- fs/fscrypt.sh none ffsb
- fs/fscrypt.sh unlocked ffsb
| 136 | 16.125 | 37 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/tasks/1-tests/fscrypt-iozone.yaml | tasks:
- workunit:
timeout: 6h
clients:
client.0:
- fs/fscrypt.sh none iozone
- fs/fscrypt.sh unlocked iozone
| 140 | 16.625 | 39 | yaml |
null | ceph-main/qa/suites/fs/fscrypt/tasks/1-tests/fscrypt-pjd.yaml | tasks:
- workunit:
timeout: 6h
clients:
client.0:
- fs/fscrypt.sh none pjd
- fs/fscrypt.sh unlocked pjd
| 134 | 15.875 | 36 | yaml |
null | ceph-main/qa/suites/fs/full/overrides.yaml | overrides:
ceph:
conf:
mgr:
debug client: 20
log-ignorelist:
- OSD full dropping all updates
- OSD near full
- pausewr flag
- failsafe engaged, dropping updates
- failsafe disengaged, no longer dropping
- is full \(reached quota
- POOL_FULL
- POOL_NEARFULL
- POOL_BACKFILLFULL
- PG_DEGRADED
- OSD_OUT_OF_ORDER_FULL
- OSD_NEARFULL
- OSD_FULL
| 443 | 21.2 | 47 | yaml |
null | ceph-main/qa/suites/fs/full/clusters/1-node-1-mds-1-osd.yaml | .qa/cephfs/clusters/1-node-1-mds-1-osd.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/full/mount/fuse.yaml | tasks:
- ceph-fuse:
| 22 | 6.666667 | 14 | yaml |
null | ceph-main/qa/suites/fs/full/objectstore/bluestore-bitmap.yaml | .qa/objectstore/bluestore-bitmap.yaml | 37 | 37 | 37 | yaml |
null | ceph-main/qa/suites/fs/full/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/full/overrides/ignorelist_wrongly_marked_down.yaml | .qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml | 56 | 56 | 56 | yaml |
null | ceph-main/qa/suites/fs/full/overrides/no_client_pidfile.yaml | .qa/overrides/no_client_pidfile.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/fs/full/tasks/mgr-osd-full.yaml | overrides:
ceph:
conf:
global:
osd_pool_default_size: 1
osd_pool_default_min_size: 1
client:
debug ms: 1
debug client: 20
mds:
debug ms: 1
debug mds: 20
osd: # force bluestore since it's required for ec overwrites
osd objectstore: bluestore
bluestore block size: 1073741824
tasks:
- workunit:
cleanup: true
clients:
client.0:
- fs/full/subvolume_rm.sh
- workunit:
cleanup: true
clients:
client.0:
- fs/full/subvolume_clone.sh
- workunit:
cleanup: true
clients:
client.0:
- fs/full/subvolume_snapshot_rm.sh
| 666 | 19.84375 | 66 | yaml |
null | ceph-main/qa/suites/fs/functional/clusters/1a3s-mds-4c-client.yaml | .qa/cephfs/clusters/1a3s-mds-4c-client.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/functional/objectstore/bluestore-bitmap.yaml | .qa/objectstore/bluestore-bitmap.yaml | 37 | 37 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/objectstore/bluestore-ec-root.yaml | .qa/cephfs/objectstore-ec/bluestore-ec-root.yaml | 48 | 48 | 48 | yaml |
null | ceph-main/qa/suites/fs/functional/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/functional/overrides/ignorelist_wrongly_marked_down.yaml | .qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml | 56 | 56 | 56 | yaml |
null | ceph-main/qa/suites/fs/functional/overrides/no_client_pidfile.yaml | .qa/overrides/no_client_pidfile.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/acls.yaml | overrides:
ceph:
log-ignorelist:
- Reduced data availability
- Degraded data redundancy
tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_acls
| 211 | 18.272727 | 33 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/admin.yaml | overrides:
ceph:
conf:
global:
lockdep: true
log-ignorelist:
- missing required features
tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_admin
| 225 | 16.384615 | 33 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/alternate-pool.yaml | overrides:
ceph:
log-ignorelist:
- bad backtrace
- object missing on disk
- error reading table object
- error reading sessionmap
- unmatched fragstat
- unmatched rstat
- was unreadable, recreating it now
- Scrub error on inode
- Metadata damage detected
- MDS_FAILED
- MDS_DAMAGE
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_recovery_pool
| 442 | 21.15 | 41 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/asok_dump_tree.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_dump_tree
| 78 | 14.8 | 35 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/auto-repair.yaml | overrides:
ceph:
log-ignorelist:
- force file system read-only
- bad backtrace
- MDS in read-only mode
- \(MDS_READ_ONLY\)
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_auto_repair
| 239 | 19 | 39 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/backtrace.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_backtrace
| 84 | 16 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/cap-flush.yaml | overrides:
ceph:
log-ignorelist:
- Replacing daemon mds.a
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_cap_flush
| 154 | 16.222222 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/client-limits.yaml | overrides:
ceph:
log-ignorelist:
- responding to mclientcaps\(revoke\)
- not advance its oldest_client_tid
- failing to advance its oldest client/flush tid
- Too many inodes in cache
- failing to respond to cache pressure
- slow requests are blocked
- failing to respond to capability release
- MDS cache is too large
- \(MDS_CLIENT_OLDEST_TID\)
- \(MDS_CACHE_OVERSIZED\)
tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_client_limits
| 552 | 28.105263 | 54 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/client-readahead.yaml | tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_readahead
| 110 | 17.5 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/client-recovery.yaml | # The task interferes with the network, so we need
# to permit OSDs to complain about that.
overrides:
ceph:
log-ignorelist:
- evicting unresponsive client
- but it is still running
- slow request
- MDS_CLIENT_LATE_RELEASE
- t responding to mclientcaps
- file system flag refuse_client_session is set
- Degraded data redundancy
- MDS_CLIENTS_LAGGY
- Reduced data availability
tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_client_recovery
| 552 | 26.65 | 53 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/damage.yaml | overrides:
ceph:
log-ignorelist:
- bad backtrace
- object missing on disk
- error reading table object
- error reading sessionmap
- Error loading MDS rank
- missing journal object
- Error recovering journal
- error decoding table object
- failed to read JournalPointer
- Corrupt directory entry
- Corrupt fnode header
- corrupt sessionmap header
- Corrupt dentry
- Scrub error on inode
- Metadata damage detected
- MDS_READ_ONLY
- force file system read-only
- with standby daemon mds
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_damage
| 681 | 24.259259 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/data-scan.yaml | overrides:
ceph:
log-ignorelist:
- bad backtrace
- object missing on disk
- error reading table object
- error reading sessionmap
- unmatched fragstat
- unmatched rstat
- was unreadable, recreating it now
- Scrub error on inode
- Metadata damage detected
- inconsistent rstat on inode
- Error recovering journal
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_data_scan
| 468 | 23.684211 | 41 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/exports.yaml | overrides:
ceph:
log-ignorelist:
- Replacing daemon mds
tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_exports
| 168 | 15.9 | 33 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/forward-scrub.yaml | overrides:
ceph:
log-ignorelist:
- inode wrongly marked free
- bad backtrace on inode
- inode table repaired for inode
- Scrub error on inode
- Scrub error on dir
- Metadata damage detected
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_forward_scrub
| 320 | 21.928571 | 41 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/fragment.yaml |
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_fragment
| 84 | 13.166667 | 36 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/journal-repair.yaml | overrides:
ceph:
log-ignorelist:
- bad backtrace on directory inode
- error reading table object
- Metadata damage detected
- slow requests are blocked
- Behind on trimming
- error reading sessionmap
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_journal_repair
| 331 | 22.714286 | 42 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/mds-flush.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_flush
| 80 | 15.2 | 33 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/mds-full.yaml | overrides:
ceph:
cephfs:
ec_profile:
- disabled
log-ignorelist:
- OSD full dropping all updates
- OSD near full
- pausewr flag
- failsafe engaged, dropping updates
- failsafe disengaged, no longer dropping
- is full \(reached quota
- POOL_FULL
- POOL_BACKFILLFULL
- PG_RECOVERY_FULL
- PG_DEGRADED
conf:
mon:
mon osd nearfull ratio: 0.6
mon osd backfillfull ratio: 0.6
mon osd full ratio: 0.7
osd:
osd mon report interval: 5
osd objectstore: memstore
osd failsafe full ratio: 1.0
memstore device bytes: 200000000
client.0:
debug client: 20
debug objecter: 20
debug objectcacher: 20
client.1:
debug client: 20
debug objecter: 20
debug objectcacher: 20
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_full
| 945 | 23.25641 | 47 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/mds_creation_retry.yaml | tasks:
-mds_creation_failure:
- workunit:
clients:
all: [fs/misc/trivial_sync.sh]
| 93 | 12.428571 | 36 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/metrics.yaml | tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_mds_metrics
| 104 | 16.5 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/multimds_misc.yaml | overrides:
ceph:
log-ignorelist:
- Scrub error on inode
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_multimds_misc
| 151 | 14.2 | 39 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/openfiletable.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_openfiletable
| 88 | 16.8 | 41 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/pool-perm.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_pool_perm
| 84 | 16 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/quota.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_quota
| 80 | 15.2 | 33 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/recovery-fs.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_recovery_fs
| 80 | 15.2 | 37 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/scrub.yaml | overrides:
ceph:
log-ignorelist:
- Replacing daemon mds
- Scrub error on inode
- Behind on trimming
- Metadata damage detected
- bad backtrace on inode
- overall HEALTH_
- \(MDS_TRIM\)
conf:
mds:
mds log max segments: 1
mds cache max size: 1000
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_scrub_checks
- tasks.cephfs.test_scrub
| 432 | 20.65 | 38 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/sessionmap.yaml | overrides:
ceph:
log-ignorelist:
- client session with non-allowable root
tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_sessionmap
| 197 | 18.8 | 46 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/snap-schedule.yaml | overrides:
ceph:
conf:
mgr:
debug mgr: 20
debug ms: 1
debug finisher: 20
debug client: 20
log-whitelist:
- OSD full dropping all updates
- OSD near full
- pausewr flag
- failsafe engaged, dropping updates
- failsafe disengaged, no longer dropping
- is full \(reached quota
- POOL_FULL
- POOL_BACKFILLFULL
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_snap_schedules
| 491 | 20.391304 | 47 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/snap_schedule_snapdir.yaml | overrides:
ceph:
conf:
mgr:
debug mgr: 20
debug ms: 1
debug finisher: 20
debug client: 20
log-whitelist:
- OSD full dropping all updates
- OSD near full
- pausewr flag
- failsafe engaged, dropping updates
- failsafe disengaged, no longer dropping
- is full \(reached quota
- POOL_FULL
- POOL_BACKFILLFULL
overrides:
kclient:
snapdirname: .customsnapkernel
ceph:
conf:
client:
client snapdir: .customsnapfuse
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_snap_schedules.TestSnapSchedulesSnapdir
| 646 | 19.870968 | 67 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/snapshots.yaml | overrides:
check-counter:
dry_run: true
ceph:
log-ignorelist:
- evicting unresponsive client
- RECENT_CRASH
tasks:
- exec:
mon.a:
- ceph config set mgr mgr/crash/warn_recent_interval 0
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.test_snapshots
| 315 | 17.588235 | 60 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/strays.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_strays
| 81 | 15.4 | 34 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/test_journal_migration.yaml | tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_journal_migration
| 86 | 16.4 | 43 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/truncate_delay.yaml | overrides:
ceph:
conf:
client:
ms_inject_delay_probability: 1
ms_inject_delay_type: osd
ms_inject_delay_max: 5
client_oc_max_dirty_age: 1
tasks:
- exec:
client.0:
- cd $TESTDIR/mnt.0 && dd if=/dev/zero of=./foo count=100
- sleep 2
- cd $TESTDIR/mnt.0 && truncate --size 0 ./foo
| 344 | 22 | 63 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/xfstests-dev.yaml | tasks:
- cephfs_test_runner:
fail_on_skip: false
modules:
- tasks.cephfs.tests_from_xfstests_dev
| 119 | 19 | 46 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/workunit/dir-max-entries.yaml | tasks:
- workunit:
clients:
client.0:
- fs/maxentries
| 72 | 11.166667 | 23 | yaml |
null | ceph-main/qa/suites/fs/functional/tasks/workunit/quota.yaml | tasks:
- workunit:
clients:
all:
- fs/quota
| 62 | 9.5 | 18 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/clusters/1-mds-1-client-coloc.yaml | .qa/cephfs/clusters/1-mds-1-client-coloc.yaml | 45 | 45 | 45 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/objectstore/bluestore-bitmap.yaml | .qa/objectstore/bluestore-bitmap.yaml | 37 | 37 | 37 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/overrides/ignorelist_wrongly_marked_down.yaml | .qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml | 56 | 56 | 56 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/overrides/no_client_pidfile.yaml | .qa/overrides/no_client_pidfile.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/tasks/client.yaml | overrides:
ceph:
conf:
client:
debug ms: 1
debug client: 20
mds:
debug ms: 1
debug mds: 20
tasks:
- workunit:
clients:
client.0:
- client/test.sh
| 214 | 13.333333 | 24 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/tasks/ino_release_cb.yaml | overrides:
ceph:
conf:
client:
debug ms: 1
debug client: 20
mds:
debug ms: 1
debug mds: 20
tasks:
- exec:
mon.a:
- ceph config set mds mds_min_caps_per_client 1
- background_exec:
mon.a:
- "sleep 30 && ceph config set mds mds_max_caps_per_client 1"
- exec:
client.0:
- ceph_test_ino_release_cb
| 368 | 17.45 | 65 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/tasks/libcephfs_python.yaml | overrides:
ceph-fuse:
disabled: true
kclient:
disabled: true
tasks:
- workunit:
clients:
client.0:
- fs/test_python.sh
| 149 | 12.636364 | 27 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/tasks/libcephfs/frag.yaml | .qa/cephfs/overrides/frag.yaml | 30 | 30 | 30 | yaml |
null | ceph-main/qa/suites/fs/libcephfs/tasks/libcephfs/test.yaml | overrides:
ceph:
conf:
client:
debug ms: 1
debug client: 20
mds:
debug ms: 1
debug mds: 20
tasks:
- check-counter:
counters:
mds:
- "mds.dir_split"
- workunit:
clients:
client.0:
- libcephfs/test.sh
| 285 | 14.052632 | 27 | yaml |
null | ceph-main/qa/suites/fs/mirror-ha/cephfs-mirror/three-per-cluster.yaml | meta:
- desc: run one cephfs-mirror daemon on primary cluster
tasks:
- cephfs-mirror:
client: client.mirror1
run_in_foreground: True
- cephfs-mirror:
client: client.mirror2
run_in_foreground: True
- cephfs-mirror:
client: client.mirror3
run_in_foreground: True
| 285 | 21 | 55 | yaml |
null | ceph-main/qa/suites/fs/mirror-ha/clients/mirror.yaml | meta:
- desc: configure the permissions for client.mirror
overrides:
ceph:
conf:
client:
debug cephfs_mirror: 20
log to stderr: false
# make these predictable
client.mirror1:
admin socket: /var/run/ceph/cephfs-mirror1.asok
pid file: /var/run/ceph/cephfs-mirror1.pid
client.mirror2:
admin socket: /var/run/ceph/cephfs-mirror2.asok
pid file: /var/run/ceph/cephfs-mirror2.pid
client.mirror3:
admin socket: /var/run/ceph/cephfs-mirror3.asok
pid file: /var/run/ceph/cephfs-mirror3.pid
tasks:
- exec:
client.mirror1:
- "sudo ceph auth caps client.mirror1 mon 'profile cephfs-mirror' mds 'allow r' osd 'allow rw tag cephfs metadata=*, allow r tag cephfs data=*' mgr 'allow r'"
client.mirror2:
- "sudo ceph auth caps client.mirror2 mon 'profile cephfs-mirror' mds 'allow r' osd 'allow rw tag cephfs metadata=*, allow r tag cephfs data=*' mgr 'allow r'"
client.mirror3:
- "sudo ceph auth caps client.mirror3 mon 'profile cephfs-mirror' mds 'allow r' osd 'allow rw tag cephfs metadata=*, allow r tag cephfs data=*' mgr 'allow r'"
client.mirror_remote:
- "sudo ceph auth caps client.mirror_remote mon 'allow r' mds 'allow rwps' osd 'allow rw tag cephfs *=*' mgr 'allow r'"
client.1:
- "sudo ceph auth caps client.0 mon 'allow r' mds 'allow rwps' osd 'allow rw tag cephfs *=*' mgr 'allow r'"
client.2:
- "sudo ceph auth caps client.1 mon 'allow r' mds 'allow rwps' osd 'allow rw tag cephfs *=*' mgr 'allow r'"
| 1,558 | 46.242424 | 164 | yaml |
null | ceph-main/qa/suites/fs/mirror-ha/cluster/1-node.yaml | meta:
- desc: 1 ceph cluster with 1 mon, 1 mgr, 3 osds, 5 mdss
roles:
- - mon.a
- mgr.x
- mds.a
- mds.b
- mds.c
- mds.d
- mds.e
- osd.0
- osd.1
- osd.2
- client.0
- client.1
- client.2
- client.mirror1
- client.mirror2
- client.mirror3
- client.mirror_remote | 290 | 13.55 | 56 | yaml |
null | ceph-main/qa/suites/fs/mirror-ha/objectstore/bluestore-bitmap.yaml | .qa/objectstore/bluestore-bitmap.yaml | 37 | 37 | 37 | yaml |
null | ceph-main/qa/suites/fs/mirror-ha/overrides/whitelist_health.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(FS_DEGRADED\)
- \(MDS_FAILED\)
- \(MDS_DEGRADED\)
- \(FS_WITH_FAILED_MDS\)
- \(MDS_DAMAGE\)
- \(MDS_ALL_DOWN\)
- \(MDS_UP_LESS_THAN_MAX\)
- \(FS_INLINE_DATA_DEPRECATED\)
- Reduced data availability
- Degraded data redundancy
| 352 | 22.533333 | 37 | yaml |
null | ceph-main/qa/suites/fs/mirror-ha/workloads/cephfs-mirror-ha-workunit.yaml | meta:
- desc: run the cephfs_mirror_ha.sh workunit to test cephfs-mirror daemon in HA active/active mode
overrides:
ceph:
conf:
mgr:
debug client: 10
tasks:
- exec:
mon.a:
- "ceph fs volume create dc"
- "ceph fs volume create dc-backup"
# Remove volumes during unwind to avoid MDS replacement warnings:
- full_sequential_finally:
- exec:
mon.a:
- ceph config set mon mon_allow_pool_delete true
- ceph fs volume rm dc --yes-i-really-mean-it
- ceph fs volume rm dc-backup --yes-i-really-mean-it
- ceph-fuse:
client.1:
cephfs_name: dc
client.2:
cephfs_name: dc-backup
- cephfs_mirror_thrash:
randomize: False
max_thrash_delay: 10
- workunit:
subdir: mirror
cleanup: False
clients:
client.1: [fs/cephfs_mirror_ha_gen.sh]
timeout: 1h
- exec:
client.2:
- "echo verifying synchronized snapshots..."
- workunit:
subdir: mirror
cleanup: False
clients:
client.2: [fs/cephfs_mirror_ha_verify.sh]
timeout: 3h | 1,116 | 24.386364 | 98 | yaml |
null | ceph-main/qa/suites/fs/mirror/cephfs-mirror/one-per-cluster.yaml | meta:
- desc: run one cephfs-mirror daemon on primary cluster
tasks:
- cephfs-mirror:
client: client.mirror
| 112 | 17.833333 | 55 | yaml |
null | ceph-main/qa/suites/fs/mirror/clients/mirror.yaml | meta:
- desc: configure the permissions for client.mirror
overrides:
ceph:
conf:
client:
debug cephfs_mirror: 20
log to stderr: false
# make these predictable
client.mirror:
admin socket: /var/run/ceph/cephfs-mirror.asok
pid file: /var/run/ceph/cephfs-mirror.pid
tasks:
- exec:
client.mirror:
- "sudo ceph auth caps client.mirror mon 'profile cephfs-mirror' mds 'allow r' osd 'allow rw tag cephfs metadata=*, allow r tag cephfs data=*' mgr 'allow r'"
client.mirror_remote:
- "sudo ceph auth caps client.mirror_remote mon 'allow r' mds 'allow rwps' osd 'allow rw tag cephfs *=*' mgr 'allow r'"
| 669 | 34.263158 | 163 | yaml |
null | ceph-main/qa/suites/fs/mirror/cluster/1-node.yaml | meta:
- desc: 1 ceph cluster with 1 mon, 1 mgr, 3 osds, 5 mdss
roles:
- - mon.a
- mgr.x
- mds.a
- mds.b
- mds.c
- mds.d
- mds.e
- osd.0
- osd.1
- osd.2
- client.0
- client.1
- client.mirror
- client.mirror_remote | 238 | 13.058824 | 56 | yaml |
null | ceph-main/qa/suites/fs/mirror/mount/fuse.yaml | tasks:
- ceph-fuse: [client.0, client.1]
| 43 | 13.666667 | 35 | yaml |
null | ceph-main/qa/suites/fs/mirror/objectstore/bluestore-bitmap.yaml | .qa/objectstore/bluestore-bitmap.yaml | 37 | 37 | 37 | yaml |
null | ceph-main/qa/suites/fs/mirror/overrides/whitelist_health.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(FS_DEGRADED\)
- \(MDS_FAILED\)
- \(MDS_DEGRADED\)
- \(FS_WITH_FAILED_MDS\)
- \(MDS_DAMAGE\)
- \(MDS_ALL_DOWN\)
- \(MDS_UP_LESS_THAN_MAX\)
- \(FS_INLINE_DATA_DEPRECATED\)
- Reduced data availability
- Degraded data redundancy
| 352 | 22.533333 | 37 | yaml |
null | ceph-main/qa/suites/fs/mirror/tasks/mirror.yaml | overrides:
ceph:
conf:
mgr:
debug client: 10
tasks:
- cephfs_test_runner:
modules:
- tasks.cephfs.test_mirroring.TestMirroring
| 164 | 14 | 51 | yaml |
null | ceph-main/qa/suites/fs/mixed-clients/clusters/1a3s-mds-2c-client.yaml | .qa/cephfs/clusters/1a3s-mds-2c-client.yaml | 43 | 43 | 43 | yaml |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.