Search is not available for this dataset
repo
stringlengths 2
152
⌀ | file
stringlengths 15
239
| code
stringlengths 0
58.4M
| file_length
int64 0
58.4M
| avg_line_length
float64 0
1.81M
| max_line_length
int64 0
12.7M
| extension_type
stringclasses 364
values |
---|---|---|---|---|---|---|
null | ceph-main/qa/suites/ceph-deploy/config/ceph_volume_bluestore.yaml | overrides:
ceph-deploy:
use-ceph-volume: True
bluestore: True
conf:
osd:
bluestore fsck on mount: true
| 142 | 16.875 | 39 | yaml |
null | ceph-main/qa/suites/ceph-deploy/config/ceph_volume_bluestore_dmcrypt.yaml | overrides:
ceph-deploy:
use-ceph-volume: True
bluestore: True
dmcrypt: True
conf:
osd:
bluestore fsck on mount: true
| 162 | 17.111111 | 39 | yaml |
null | ceph-main/qa/suites/ceph-deploy/config/ceph_volume_dmcrypt_off.yaml | overrides:
ceph-deploy:
use-ceph-volume: True
| 55 | 13 | 27 | yaml |
null | ceph-main/qa/suites/ceph-deploy/distros/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/ceph-deploy/distros/ubuntu_latest.yaml | .qa/distros/supported/ubuntu_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/ceph-deploy/python_versions/python_2.yaml | overrides:
ceph-deploy:
python_version: "2"
| 50 | 11.75 | 23 | yaml |
null | ceph-main/qa/suites/ceph-deploy/python_versions/python_3.yaml | overrides:
ceph-deploy:
python_version: "3"
| 50 | 11.75 | 23 | yaml |
null | ceph-main/qa/suites/ceph-deploy/tasks/ceph-admin-commands.yaml | meta:
- desc: "test basic ceph admin commands"
tasks:
- ssh_keys:
- print: "**** done ssh_keys"
- ceph-deploy:
- print: "**** done ceph-deploy"
- workunit:
clients:
client.0:
- ceph-tests/ceph-admin-commands.sh
- print: "**** done ceph-tests/ceph-admin-commands.sh"
| 290 | 21.384615 | 54 | yaml |
null | ceph-main/qa/suites/ceph-deploy/tasks/rbd_import_export.yaml | meta:
- desc: "Setup cluster using ceph-deploy, Run the rbd import/export tests"
tasks:
- ssh-keys:
- ceph-deploy:
- workunit:
clients:
client.0:
- rbd/import_export.sh
| 187 | 17.8 | 74 | yaml |
null | ceph-main/qa/suites/cephmetrics/0-clusters/3-node.yaml | meta:
- desc: "4-node cluster"
roles:
- [mon.a, mds.a, osd.0, osd.1, osd.2]
- [mon.b, mgr.x, osd.3, osd.4, osd.5]
- [mon.c, mgr.y, osd.6, osd.7, osd.8, client.0]
- [cephmetrics.0]
openstack:
- volumes: # attached to each instance
count: 3
size: 10 # GB
| 261 | 20.833333 | 47 | yaml |
null | ceph-main/qa/suites/cephmetrics/1-distros/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/cephmetrics/1-distros/ubuntu_latest.yaml | .qa/distros/supported/ubuntu_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/cephmetrics/2-ceph/ceph_ansible.yaml | meta:
- desc: "Build the ceph cluster using ceph-ansible"
overrides:
ceph_ansible:
vars:
ceph_conf_overrides:
global:
osd default pool size: 2
mon pg warn min per osd: 2
osd pool default pg num: 64
osd pool default pgp num: 64
mon_max_pg_per_osd: 1024
ceph_test: true
ceph_stable_release: luminous
osd_scenario: collocated
journal_size: 1024
osd_auto_discovery: false
ceph_origin: repository
ceph_repository: dev
ceph_mgr_modules:
- status
- restful
cephfs_pools:
- name: "cephfs_data"
pg_num: "64"
- name: "cephfs_metadata"
pg_num: "64"
tasks:
- ssh-keys:
- ceph_ansible:
- install.ship_utilities:
| 822 | 23.939394 | 51 | yaml |
null | ceph-main/qa/suites/cephmetrics/3-ceph-config/bluestore_with_dmcrypt.yaml | meta:
- desc: "use bluestore + dmcrypt"
overrides:
ceph_ansible:
vars:
osd_objectstore: bluestore
dmcrypt: True
| 137 | 14.333333 | 34 | yaml |
null | ceph-main/qa/suites/cephmetrics/3-ceph-config/bluestore_without_dmcrypt.yaml | meta:
- desc: "use bluestore without dmcrypt"
overrides:
ceph_ansible:
vars:
osd_objectstore: bluestore
dmcrypt: False
| 144 | 15.111111 | 39 | yaml |
null | ceph-main/qa/suites/cephmetrics/3-ceph-config/dmcrypt_off.yaml | meta:
- desc: "without dmcrypt"
overrides:
ceph_ansible:
vars:
dmcrypt: False
| 95 | 11 | 25 | yaml |
null | ceph-main/qa/suites/cephmetrics/3-ceph-config/dmcrypt_on.yaml | meta:
- desc: "with dmcrypt"
overrides:
ceph_ansible:
vars:
dmcrypt: True
| 91 | 10.5 | 22 | yaml |
null | ceph-main/qa/suites/cephmetrics/4-epel/no_epel.yaml | meta:
- desc: "Without EPEL"
overrides:
cephmetrics:
group_vars:
all:
use_epel: false
| 108 | 12.625 | 24 | yaml |
null | ceph-main/qa/suites/cephmetrics/4-epel/use_epel.yaml | meta:
- desc: "Using EPEL"
overrides:
cephmetrics:
group_vars:
all:
use_epel: true
| 105 | 12.25 | 22 | yaml |
null | ceph-main/qa/suites/cephmetrics/5-containers/containerized.yaml | meta:
- desc: "Containerized prometheus and grafana"
overrides:
cephmetrics:
group_vars:
all:
prometheus:
containerized: true
grafana:
containerized: true
| 205 | 17.727273 | 48 | yaml |
null | ceph-main/qa/suites/cephmetrics/5-containers/no_containers.yaml | meta:
- desc: "Packaged prometheus and grafana"
overrides:
cephmetrics:
group_vars:
all:
prometheus:
containerized: false
grafana:
containerized: false
| 202 | 17.454545 | 43 | yaml |
null | ceph-main/qa/suites/cephmetrics/6-tasks/cephmetrics.yaml | meta:
- desc: "Deploy cephmetrics and run integration tests"
tasks:
- cephmetrics:
| 83 | 15.8 | 54 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/crimson_qa_overrides.yaml | .qa/config/crimson_qa_overrides.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/clusters/fixed-1.yaml | overrides:
ceph-deploy:
conf:
global:
osd pool default size: 2
osd crush chooseleaf type: 0
osd pool default pg num: 128
osd pool default pgp num: 128
ceph:
conf:
osd:
osd shutdown pgref assert: true
roles:
- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0]
| 321 | 20.466667 | 47 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/clusters/fixed-2.yaml | roles:
- [mon.a, osd.0, osd.1, client.0, node-exporter.a]
- [mgr.x, osd.2, osd.3, client.1, prometheus.a, node-exporter.b]
overrides:
ceph:
conf:
osd:
osd shutdown pgref assert: true
global:
ms cluster mode: crc
ms service mode: crc
ms client mode: crc
ms mon service mode: crc
ms mon cluster mode: crc
ms mon client mode: crc
| 401 | 24.125 | 64 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/deploy/ceph.yaml | overrides:
install:
ceph:
flavor: crimson
tasks:
- install:
- ceph:
conf:
osd:
debug monc: 20
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
flavor: crimson
| 383 | 19.210526 | 41 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/objectstore/seastore.yaml | overrides:
ceph:
fs: xfs
conf:
osd:
osd objectstore: seastore
| 86 | 11.428571 | 33 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/tasks/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- but it is still running
- overall HEALTH_
- \(POOL_FULL\)
- \(SMALLER_PGP_NUM\)
- \(CACHE_POOL_NO_HIT_SET\)
- \(CACHE_POOL_NEAR_FULL\)
- \(POOL_APP_NOT_ENABLED\)
- \(PG_AVAILABILITY\)
- \(PG_DEGRADED\)
conf:
client:
debug ms: 1
mon:
mon warn on pool no app: false
osd:
osd class load list: "*"
osd class default list: "*"
osd blocked scrub grace period: 3600
tasks:
- workunit:
clients:
client.0:
- rados/test.sh
- rados/test_pool_quota.sh
| 625 | 20.586207 | 44 | yaml |
null | ceph-main/qa/suites/crimson-rados-experimental/seastore/basic/tasks/readwrite.yaml | overrides:
ceph:
crush_tunables: optimal
conf:
mon:
mon osd initial require min compat client: luminous
osd:
osd_discard_disconnected_ops: false
tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
op_weights:
read: 45
write: 45
delete: 10
| 318 | 16.722222 | 59 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/crimson_qa_overrides.yaml | .qa/config/crimson_qa_overrides.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/clusters/fixed-2.yaml | roles:
- [mon.a, osd.0, osd.1, client.0, node-exporter.a]
- [mgr.x, osd.2, osd.3, client.1, prometheus.a, node-exporter.b]
overrides:
ceph:
conf:
osd:
osd shutdown pgref assert: true
global:
ms cluster mode: crc
ms service mode: crc
ms client mode: crc
ms mon service mode: crc
ms mon cluster mode: crc
ms mon client mode: crc
| 401 | 24.125 | 64 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/deploy/ceph.yaml | overrides:
install:
ceph:
flavor: crimson
tasks:
- install:
- ceph:
conf:
osd:
debug monc: 20
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
flavor: crimson
| 383 | 19.210526 | 41 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/tasks/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- but it is still running
- overall HEALTH_
- \(POOL_FULL\)
- \(SMALLER_PGP_NUM\)
- \(CACHE_POOL_NO_HIT_SET\)
- \(CACHE_POOL_NEAR_FULL\)
- \(POOL_APP_NOT_ENABLED\)
- \(PG_AVAILABILITY\)
- \(PG_DEGRADED\)
conf:
client:
debug ms: 1
mon:
mon warn on pool no app: false
osd:
osd class load list: "*"
osd class default list: "*"
osd blocked scrub grace period: 3600
tasks:
- workunit:
clients:
client.0:
- rados/test.sh
- rados/test_pool_quota.sh
| 625 | 20.586207 | 44 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/tasks/rados_python.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- overall HEALTH_
- \(OSDMAP_FLAGS\)
- \(PG_
- \(OSD_
- \(OBJECT_
- \(POOL_APP_NOT_ENABLED\)
tasks:
- workunit:
timeout: 1h
clients:
client.0:
- rados/test_python.sh --eval-attr 'not (wait or tier or ec or bench or stats)'
| 338 | 18.941176 | 87 | yaml |
null | ceph-main/qa/suites/crimson-rados/basic/tasks/readwrite.yaml | overrides:
ceph:
crush_tunables: optimal
conf:
mon:
mon osd initial require min compat client: luminous
osd:
osd_discard_disconnected_ops: false
tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
op_weights:
read: 45
write: 45
delete: 10
| 318 | 16.722222 | 59 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/crimson_qa_overrides.yaml | .qa/config/crimson_qa_overrides.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/clusters/fixed-2.yaml | roles:
- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0]
overrides:
ceph:
conf:
osd:
osd shutdown pgref assert: true
global:
ms cluster mode: crc
ms service mode: crc
ms client mode: crc
ms mon service mode: crc
ms mon cluster mode: crc
ms mon client mode: crc
| 333 | 21.266667 | 47 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/deploy/ceph.yaml | overrides:
install:
ceph:
flavor: crimson
tasks:
- install:
- ceph:
wait-for-scrub: false
conf:
osd:
debug monc: 20
flavor: crimson
| 170 | 12.153846 | 25 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/objectstore/bluestore.yaml | overrides:
ceph:
fs: xfs
conf:
osd:
osd objectstore: bluestore
bluestore block size: 96636764160
debug bluestore: 20
debug bluefs: 20
debug rocksdb: 10
bluestore compression mode: aggressive
bluestore fsck on mount: true
bluestore compression algorithm: snappy
# lower the full ratios since we can fill up a 100gb osd so quickly
mon osd full ratio: .9
mon osd backfillfull_ratio: .85
mon osd nearfull ratio: .8
osd failsafe full ratio: .95
bluestore rocksdb cf: false
log to stderr: true
err to stderr: true
log flush on exit: true
log to file: false
| 711 | 28.666667 | 75 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/settings/optimized.yaml | overrides:
ceph:
conf:
mon:
debug mon: "0/0"
debug ms: "0/0"
debug paxos: "0/0"
osd:
debug filestore: "0/0"
debug journal: "0/0"
debug ms: "0/0"
debug osd: "0/0"
global:
auth client required: none
auth cluster required: none
auth service required: none
auth supported: none
debug lockdep: "0/0"
debug context: "0/0"
debug crush: "0/0"
debug mds: "0/0"
debug mds balancer: "0/0"
debug mds locker: "0/0"
debug mds log: "0/0"
debug mds log expire: "0/0"
debug mds migrator: "0/0"
debug buffer: "0/0"
debug timer: "0/0"
debug filer: "0/0"
debug striper: "0/0"
debug objecter: "0/0"
debug rados: "0/0"
debug rbd: "0/0"
debug rbd mirror: "0/0"
debug rbd replay: "0/0"
debug journaler: "0/0"
debug objectcacher: "0/0"
debug client: "0/0"
debug osd: "0/0"
debug optracker: "0/0"
debug objclass: "0/0"
debug filestore: "0/0"
debug journal: "0/0"
debug ms: "0/0"
debug mon: "0/0"
debug monc: "0/0"
debug paxos: "0/0"
debug tp: "0/0"
debug auth: "0/0"
debug crypto: "0/0"
debug finisher: "0/0"
debug heartbeatmap: "0/0"
debug perfcounter: "0/0"
debug rgw: "0/0"
debug rgw sync: "0/0"
debug civetweb: "0/0"
debug javaclient: "0/0"
debug asok: "0/0"
debug throttle: "0/0"
debug refs: "0/0"
debug compressor: "0/0"
debug bluestore: "0/0"
debug bluefs: "0/0"
debug bdev: "0/0"
debug kstore: "0/0"
debug rocksdb: "0/0"
debug leveldb: "0/0"
debug memdb: "0/0"
debug fuse: "0/0"
debug mgr: "0/0"
debug mgrc: "0/0"
debug dpdk: "0/0"
debug eventtrace: "0/0"
| 1,999 | 25.666667 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/fio_4K_rand_read.yaml | tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4096]
time: 60
mode: ['randread']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iodepth: [32]
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
rbd:
pg_size: 128
pgp_size: 128
replication: 3
| 508 | 19.36 | 31 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/fio_4K_rand_rw.yaml | tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4096]
time: 60
mode: ['randrw']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iodepth: [32]
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
rbd:
pg_size: 128
pgp_size: 128
replication: 3
| 506 | 19.28 | 31 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/fio_4M_rand_read.yaml | tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4194304]
time: 60
mode: ['randread']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iodepth: [32]
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
rbd:
pg_size: 128
pgp_size: 128
replication: 3
| 511 | 19.48 | 31 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/fio_4M_rand_rw.yaml | tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4194304]
time: 60
mode: ['randrw']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iodepth: [32]
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
rbd:
pg_size: 128
pgp_size: 128
replication: 3
| 509 | 19.4 | 31 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/fio_4M_rand_write.yaml | tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4194304]
time: 60
mode: ['randwrite']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iodepth: [32]
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
rbd:
pg_size: 128
pgp_size: 128
replication: 3
| 512 | 19.52 | 31 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/radosbench_4K_rand_read.yaml | tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4096]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
- collectl
time: 60
write_only: false
readmode: 'rand'
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
replicated:
pg_size: 256
pgp_size: 256
replication: 'replicated'
| 527 | 20.12 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/radosbench_4K_seq_read.yaml | tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4096]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
- collectl
time: 60
write_only: false
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
replicated:
pg_size: 256
pgp_size: 256
replication: 'replicated'
| 502 | 19.958333 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/radosbench_4M_rand_read.yaml | tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4194304]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
- collectl
time: 60
write_only: false
readmode: 'rand'
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
replicated:
pg_size: 256
pgp_size: 256
replication: 'replicated'
| 530 | 20.24 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/radosbench_4M_seq_read.yaml | tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4194304]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
- collectl
time: 60
write_only: false
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
replicated:
pg_size: 256
pgp_size: 256
replication: 'replicated'
| 505 | 20.083333 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/radosbench_4M_write.yaml | tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4194304]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
- collectl
time: 60
write_only: true
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
replicated:
pg_size: 256
pgp_size: 256
replication: 'replicated'
| 504 | 20.041667 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/radosbench_omap_write.yaml | tasks:
- radosbench:
clients: [client.0]
write-omap: True
objectsize: 4096
size: 4096
time: 300
| 116 | 13.625 | 23 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/sample_fio.yaml | tasks:
- cbt:
benchmarks:
librbdfio:
op_size: [4096]
time: 60
mode: ['randwrite']
norandommap: True
vol_size: 4096
procs_per_volume: [1]
volumes_per_client: [2]
iodepth: [32]
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
rbd:
pg_size: 128
pgp_size: 128
replication: 3
| 509 | 19.4 | 31 | yaml |
null | ceph-main/qa/suites/crimson-rados/perf/workloads/sample_radosbench.yaml | tasks:
- cbt:
benchmarks:
radosbench:
concurrent_ops: 4
concurrent_procs: 2
op_size: [4096]
pool_monitoring_list:
- collectl
pool_profile: 'replicated'
run_monitoring_list:
- collectl
time: 60
write_only: true
cluster:
user: 'ubuntu'
osds_per_node: 3
iterations: 1
pool_profiles:
replicated:
pg_size: 256
pgp_size: 256
replication: 'replicated'
| 501 | 19.916667 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/crimson_qa_overrides.yaml | .qa/config/crimson_qa_overrides.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/clusters/fixed-1.yaml | overrides:
ceph-deploy:
conf:
global:
osd pool default size: 2
osd crush chooseleaf type: 0
osd pool default pg num: 128
osd pool default pgp num: 128
ceph:
conf:
osd:
osd shutdown pgref assert: true
global:
ms cluster mode: crc
ms service mode: crc
ms client mode: crc
ms mon service mode: crc
ms mon cluster mode: crc
ms mon client mode: crc
roles:
- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0]
| 519 | 22.636364 | 47 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/deploy/ceph.yaml | overrides:
install:
ceph:
flavor: crimson
tasks:
- install:
- ceph:
conf:
osd:
debug monc: 20
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
flavor: crimson
| 383 | 19.210526 | 41 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/tasks/rbd_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
- \(POOL_APP_NOT_ENABLED\)
- is full \(reached quota
- \(OSDMAP_FLAGS\)
- \(OSD_
- \(PG_
- \(POOL_
- \(CACHE_POOL_
- \(POOL_FULL\)
tasks:
- workunit:
clients:
client.0:
- rbd/crimson/test_crimson_librbd.sh
env:
RBD_FEATURES: "61"
| 404 | 17.409091 | 44 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/tasks/rbd_api_tests_old_format.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
- \(POOL_APP_NOT_ENABLED\)
- is full \(reached quota
- \(POOL_FULL\)
tasks:
- workunit:
clients:
client.0:
- rbd/crimson/test_crimson_librbd.sh
| 277 | 18.857143 | 44 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/tasks/rbd_cls_tests.yaml | tasks:
- workunit:
clients:
client.0:
- cls/test_cls_rbd.sh
- cls/test_cls_lock.sh
- cls/test_cls_journal.sh
| 143 | 17 | 33 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/tasks/rbd_lock_and_fence.yaml | tasks:
- workunit:
clients:
client.0:
- rbd/test_lock_fence.sh
| 81 | 12.666667 | 32 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/tasks/rbd_python_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- \(SLOW_OPS\)
- slow request
tasks:
- workunit:
clients:
client.0:
- rbd/test_librbd_python.sh --eval-attr 'not (SKIP_IF_CRIMSON)'
env:
RBD_FEATURES: "61"
| 235 | 17.153846 | 71 | yaml |
null | ceph-main/qa/suites/crimson-rados/rbd/tasks/rbd_python_api_tests_old_format.yaml | overrides:
ceph:
log-ignorelist:
- \(SLOW_OPS\)
- slow request
tasks:
- workunit:
clients:
client.0:
- rbd/test_librbd_python.sh --eval-attr 'not (SKIP_IF_CRIMSON)'
| 201 | 17.363636 | 71 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/crimson_qa_overrides.yaml | .qa/config/crimson_qa_overrides.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/thrashosds-health.yaml | .qa/tasks/thrashosds-health.yaml | 32 | 32 | 32 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml | .qa/overrides/3-size-2-min-size.yaml | 36 | 36 | 36 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/1-pg-log-overrides/normal_pg_log.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/crimson-rados/thrash/2-recovery-overrides/default.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/crimson-rados/thrash/clusters/fixed-2.yaml | roles:
- [mon.a, osd.0, osd.1, client.0, node-exporter.a]
- [mgr.x, osd.2, osd.3, client.1, prometheus.a, node-exporter.b]
overrides:
ceph:
conf:
osd:
osd shutdown pgref assert: true
global:
ms cluster mode: crc
ms service mode: crc
ms client mode: crc
ms mon service mode: crc
ms mon cluster mode: crc
ms mon client mode: crc
| 401 | 24.125 | 64 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/deploy/ceph.yaml | overrides:
install:
ceph:
flavor: crimson
tasks:
- install:
- ceph:
conf:
osd:
debug monc: 20
flavor: crimson
| 144 | 11.083333 | 22 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/objectstore/bluestore.yaml | overrides:
ceph:
fs: xfs
conf:
osd:
osd objectstore: bluestore
bluestore block size: 96636764160
debug bluestore: 20
debug bluefs: 20
debug rocksdb: 10
bluestore compression mode: aggressive
bluestore fsck on mount: true
bluestore compression algorithm: snappy
# lower the full ratios since we can fill up a 100gb osd so quickly
mon osd full ratio: .9
mon osd backfillfull_ratio: .85
mon osd nearfull ratio: .8
osd failsafe full ratio: .95
bluestore rocksdb cf: false
log to stderr: true
err to stderr: true
log flush on exit: true
log to file: false
| 711 | 28.666667 | 75 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/thrashers/default.yaml | overrides:
ceph:
wait-for-scrub: false
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
conf:
osd:
osd debug reject backfill probability: .3
osd scrub min interval: 60
osd scrub max interval: 120
osd max backfills: 3
osd snap trim sleep: 2
osd delete sleep: 1
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
tasks:
- thrashosds:
timeout: 2400
dump_ops_enable: false
sighup_delay: 0
min_in: 3
noscrub_toggle_delay: 0
chance_down: 0
chance_thrash_pg_upmap: 0
reweight_osd: 0
thrash_primary_affinity: false
ceph_objectstore_tool: false
chance_inject_pause_short: 0
chance_thrash_cluster_full: 0
| 938 | 25.083333 | 49 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/admin_socket_objecter_requests.yaml | overrides:
ceph:
conf:
client.0:
admin socket: /var/run/ceph/ceph-$name.asok
tasks:
- radosbench:
clients: [client.0]
time: 150
- admin_socket:
client.0:
objecter_requests:
test: "http://git.ceph.com/?p={repo};a=blob_plain;f=src/test/admin_socket/objecter_requests;hb={branch}"
| 324 | 22.214286 | 112 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/pool-snaps-few-objects.yaml | overrides:
conf:
osd:
osd deep scrub update digest min age: 0
tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
pool_snaps: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
| 349 | 15.666667 | 45 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/radosbench-high-concurrency.yaml | overrides:
ceph:
conf:
client.0:
debug ms: 1
debug objecter: 20
debug rados: 20
tasks:
- full_sequential:
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
| 918 | 17.38 | 26 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/radosbench.yaml | overrides:
ceph:
conf:
client.0:
debug ms: 1
debug objecter: 20
debug rados: 20
tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
| 427 | 16.12 | 26 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/small-objects-balanced.yaml | overrides:
ceph:
crush_tunables: jewel
tasks:
- rados:
clients: [client.0]
ops: 400000
max_seconds: 600
max_in_flight: 64
objects: 1024
size: 16384
balance_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
setattr: 25
rmattr: 25
| 420 | 16.541667 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/small-objects-localized.yaml | overrides:
ceph:
crush_tunables: jewel
tasks:
- rados:
clients: [client.0]
ops: 400000
max_seconds: 600
max_in_flight: 64
objects: 1024
size: 16384
localize_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
setattr: 25
rmattr: 25
| 421 | 16.583333 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/small-objects.yaml | overrides:
ceph:
crush_tunables: jewel
tasks:
- rados:
clients: [client.0]
ops: 400000
max_seconds: 600
max_in_flight: 64
objects: 1024
size: 16384
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
setattr: 25
rmattr: 25
| 396 | 16.26087 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/snaps-few-objects-balanced.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
balance_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
| 277 | 16.375 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/snaps-few-objects-localized.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
localize_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
| 278 | 16.4375 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/snaps-few-objects.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 0
# TODO: CEPH_OSD_OP_COPY_FROM
copy_from: 0
| 253 | 15.933333 | 35 | yaml |
null | ceph-main/qa/suites/crimson-rados/thrash/workloads/write_fadvise_dontneed.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
write_fadvise_dontneed: true
op_weights:
write: 100
| 137 | 14.333333 | 32 | yaml |
null | ceph-main/qa/suites/dummy/all/nop.yaml | roles:
- [mon.a, mgr.x, mds.a, osd.0, osd.1, client.0]
tasks:
- nop:
| 79 | 10.428571 | 51 | yaml |
null | ceph-main/qa/suites/experimental/multimds/clusters/7-multimds.yaml | roles:
- [mon.a, mgr.x, mds.a, mds.d]
- [mon.b, mgr.y, mds.b, mds.e]
- [mon.c, mgr.z, mds.c, mds.f]
- [osd.0]
- [osd.1]
- [osd.2]
- [client.0]
| 143 | 15 | 30 | yaml |
null | ceph-main/qa/suites/experimental/multimds/tasks/fsstress_thrash_subtrees.yaml | tasks:
- install:
- ceph:
conf:
mds:
mds thrash exports: 1
mds debug subtrees: 1
mds debug scatterstat: 1
mds verify scatter: 1
- ceph-fuse:
- workunit:
clients:
client.0:
- suites/fsstress.sh
| 254 | 14.9375 | 32 | yaml |
null | ceph-main/qa/suites/fs/32bits/clusters/fixed-2-ucephfs.yaml | .qa/cephfs/clusters/fixed-2-ucephfs.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/fs/32bits/mount/fuse.yaml | .qa/cephfs/mount/fuse.yaml | 26 | 26 | 26 | yaml |
null | ceph-main/qa/suites/fs/32bits/overrides/faked-ino.yaml | overrides:
ceph:
conf:
client:
client use faked inos: true
| 79 | 12.333333 | 35 | yaml |
null | ceph-main/qa/suites/fs/32bits/overrides/ignorelist_health.yaml | .qa/cephfs/overrides/ignorelist_health.yaml | 43 | 43 | 43 | yaml |
null | ceph-main/qa/suites/fs/32bits/overrides/ignorelist_wrongly_marked_down.yaml | .qa/cephfs/overrides/ignorelist_wrongly_marked_down.yaml | 56 | 56 | 56 | yaml |
null | ceph-main/qa/suites/fs/32bits/tasks/cfuse_workunit_suites_fsstress.yaml | .qa/cephfs/tasks/cfuse_workunit_suites_fsstress.yaml | 52 | 52 | 52 | yaml |
null | ceph-main/qa/suites/fs/32bits/tasks/cfuse_workunit_suites_pjd.yaml | overrides:
ceph:
conf:
client:
fuse set user groups: true
tasks:
- workunit:
timeout: 6h
clients:
all:
- suites/pjd.sh
| 161 | 12.5 | 34 | yaml |
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/clusters/small-cluster.yaml | roles:
- [mon.a, mgr.x, osd.0, osd.1, osd.2, mds.a, mds.b, client.0]
openstack:
- volumes: # attached to each instance
count: 2
size: 10 # GB
- machine:
disk: 100 # GB
log-rotate:
ceph-mds: 10G
ceph-osd: 10G
| 224 | 17.75 | 61 | yaml |
null | ceph-main/qa/suites/fs/bugs/client_trim_caps/objectstore/bluestore-bitmap.yaml | .qa/objectstore/bluestore-bitmap.yaml | 37 | 37 | 37 | yaml |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.