Search is not available for this dataset
repo
stringlengths 2
152
⌀ | file
stringlengths 15
239
| code
stringlengths 0
58.4M
| file_length
int64 0
58.4M
| avg_line_length
float64 0
1.81M
| max_line_length
int64 0
12.7M
| extension_type
stringclasses 364
values |
---|---|---|---|---|---|---|
null | ceph-main/qa/suites/rados/thrash/2-recovery-overrides/default.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-active-recovery.yaml | .qa/overrides/more-active-recovery.yaml | 39 | 39 | 39 | yaml |
null | ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-async-partial-recovery.yaml | overrides:
ceph:
conf:
global:
osd_async_recovery_min_cost: 1
osd_object_clean_region_max_num_intervals: 1000
| 138 | 18.857143 | 55 | yaml |
null | ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-async-recovery.yaml | overrides:
ceph:
conf:
global:
osd_async_recovery_min_cost: 1
| 82 | 12.833333 | 38 | yaml |
null | ceph-main/qa/suites/rados/thrash/2-recovery-overrides/more-partial-recovery.yaml | overrides:
ceph:
conf:
global:
osd_object_clean_region_max_num_intervals: 1000
| 99 | 15.666667 | 55 | yaml |
null | ceph-main/qa/suites/rados/thrash/3-scrub-overrides/default.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-2.yaml | overrides:
ceph:
conf:
osd:
osd max scrubs: 2
| 66 | 10.166667 | 25 | yaml |
null | ceph-main/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-3.yaml | overrides:
ceph:
conf:
osd:
osd max scrubs: 3
| 66 | 10.166667 | 25 | yaml |
null | ceph-main/qa/suites/rados/thrash/backoff/normal.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/thrash/backoff/peering.yaml | overrides:
ceph:
conf:
osd:
osd backoff on peering: true
| 77 | 12 | 36 | yaml |
null | ceph-main/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml | overrides:
ceph:
conf:
osd:
osd backoff on peering: true
osd backoff on degraded: true
| 115 | 15.571429 | 37 | yaml |
null | ceph-main/qa/suites/rados/thrash/clusters/fixed-2.yaml | .qa/clusters/fixed-2.yaml | 25 | 25 | 25 | yaml |
null | ceph-main/qa/suites/rados/thrash/clusters/openstack.yaml | openstack:
- volumes: # attached to each instance
count: 4
size: 30 # GB
| 87 | 16.6 | 40 | yaml |
null | ceph-main/qa/suites/rados/thrash/crc-failures/bad_map_crc_failure.yaml | overrides:
ceph:
conf:
osd:
osd inject bad map crc probability: 0.1
log-ignorelist:
- failed to encode map
| 137 | 16.25 | 47 | yaml |
null | ceph-main/qa/suites/rados/thrash/crc-failures/default.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/thrash/d-balancer/crush-compat.yaml | tasks:
- exec:
mon.a:
- while ! ceph balancer status ; do sleep 1 ; done
- ceph balancer mode crush-compat
- ceph balancer on
| 148 | 20.285714 | 56 | yaml |
null | ceph-main/qa/suites/rados/thrash/d-balancer/on.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/thrash/msgr-failures/fastclose.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 2500
ms tcp read timeout: 5
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 208 | 19.9 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
osd:
osd heartbeat use min delay socket: true
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 237 | 20.636364 | 48 | yaml |
null | ceph-main/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 2500
ms inject delay type: osd
ms inject delay probability: .005
ms inject delay max: 1
ms inject internal delays: .002
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 324 | 24 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/msgr-failures/osd-dispatch-delay.yaml | overrides:
ceph:
conf:
global:
osd debug inject dispatch delay duration: 0.1
osd debug inject dispatch delay probability: 0.1
| 155 | 18.5 | 56 | yaml |
null | ceph-main/qa/suites/rados/thrash/thrashers/careful.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
conf:
osd:
osd debug reject backfill probability: .3
osd scrub min interval: 60
osd scrub max interval: 120
osd max backfills: 3
osd snap trim sleep: 2
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
tasks:
- thrashosds:
timeout: 1200
chance_pgnum_grow: 1
chance_pgnum_shrink: 1
chance_pgpnum_fix: 1
aggressive_pg_num_changes: false
| 705 | 25.148148 | 49 | yaml |
null | ceph-main/qa/suites/rados/thrash/thrashers/default.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
conf:
osd:
osd debug reject backfill probability: .3
osd scrub min interval: 60
osd scrub max interval: 120
osd max backfills: 3
osd snap trim sleep: 2
osd delete sleep: 1
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
tasks:
- thrashosds:
timeout: 1200
chance_pgnum_grow: 1
chance_pgnum_shrink: 1
chance_pgpnum_fix: 1
chance_bluestore_reshard: 1
bluestore_new_sharding: random
| 763 | 25.344828 | 49 | yaml |
null | ceph-main/qa/suites/rados/thrash/thrashers/mapgap.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
- osd_map_cache_size
conf:
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
osd:
osd map cache size: 1
osd scrub min interval: 60
osd scrub max interval: 120
osd scrub during recovery: false
osd max backfills: 6
tasks:
- thrashosds:
timeout: 1800
chance_pgnum_grow: 0.25
chance_pgnum_shrink: 0.25
chance_pgpnum_fix: 0.25
chance_test_map_discontinuity: 2
| 729 | 25.071429 | 41 | yaml |
null | ceph-main/qa/suites/rados/thrash/thrashers/morepggrow.yaml | overrides:
ceph:
conf:
osd:
osd scrub min interval: 60
osd scrub max interval: 120
journal throttle high multiple: 2
journal throttle max multiple: 10
filestore queue throttle high multiple: 2
filestore queue throttle max multiple: 10
osd max backfills: 9
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
tasks:
- thrashosds:
timeout: 1200
chance_pgnum_grow: 3
chance_pgpnum_fix: 1
openstack:
- volumes:
size: 50
| 540 | 22.521739 | 49 | yaml |
null | ceph-main/qa/suites/rados/thrash/thrashers/none.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/thrash/thrashers/pggrow.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
conf:
osd:
osd scrub min interval: 60
osd scrub max interval: 120
filestore odsync write: true
osd max backfills: 2
osd snap trim sleep: .5
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
tasks:
- thrashosds:
timeout: 1200
chance_pgnum_grow: 2
chance_pgpnum_fix: 1
| 629 | 24.2 | 41 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml | overrides:
ceph:
conf:
client.0:
admin socket: /var/run/ceph/ceph-$name.asok
tasks:
- radosbench:
clients: [client.0]
time: 150
- admin_socket:
client.0:
objecter_requests:
test: "http://git.ceph.com/?p={repo};a=blob_plain;f=src/test/admin_socket/objecter_requests;hb={branch}"
| 324 | 22.214286 | 112 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache-agent-big.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd erasure-code-profile set myprofile crush-failure-domain=osd m=2 k=2
- sudo ceph osd pool create base 4 4 erasure myprofile
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool set base min_size 2
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 60
- sudo ceph osd pool set cache target_max_objects 5000
- rados:
clients: [client.0]
pools: [base]
ops: 10000
objects: 6600
max_seconds: 1200
size: 1024
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
| 1,167 | 30.567568 | 89 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache-agent-small.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd pool create base 4
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 60
- sudo ceph osd pool set cache target_max_objects 250
- sudo ceph osd pool set cache min_read_recency_for_promote 2
- sudo ceph osd pool set cache min_write_recency_for_promote 2
- rados:
clients: [client.0]
pools: [base]
ops: 4000
objects: 500
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
| 1,107 | 30.657143 | 78 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd pool create base 4
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache readproxy
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 3600
- sudo ceph osd pool set cache target_max_objects 250
- rados:
clients: [client.0]
pools: [base]
ops: 4000
objects: 500
pool_snaps: true
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
cache_flush: 50
cache_try_flush: 50
cache_evict: 50
snap_create: 50
snap_remove: 50
rollback: 50
| 1,126 | 27.175 | 78 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd pool create base 4
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 3600
- sudo ceph osd pool set cache target_max_objects 250
- sudo ceph osd pool set cache min_read_recency_for_promote 0
- sudo ceph osd pool set cache min_write_recency_for_promote 0
- rados:
clients: [client.0]
pools: [base]
ops: 4000
objects: 500
pool_snaps: true
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
cache_flush: 50
cache_try_flush: 50
cache_evict: 50
snap_create: 50
snap_remove: 50
rollback: 50
openstack:
- machine:
ram: 15000 # MB
| 1,309 | 28.111111 | 78 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache-snaps-balanced.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd pool create base 4
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 3600
- sudo ceph osd pool set cache target_max_objects 250
- sudo ceph osd pool set cache min_read_recency_for_promote 2
- rados:
clients: [client.0]
pools: [base]
ops: 4000
objects: 500
balance_reads: true
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
cache_flush: 50
cache_try_flush: 50
cache_evict: 50
snap_create: 50
snap_remove: 50
rollback: 50
| 1,197 | 28.219512 | 78 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache-snaps.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd pool create base 4
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 3600
- sudo ceph osd pool set cache target_max_objects 250
- sudo ceph osd pool set cache min_read_recency_for_promote 2
- rados:
clients: [client.0]
pools: [base]
ops: 4000
objects: 500
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
cache_flush: 50
cache_try_flush: 50
cache_evict: 50
snap_create: 50
snap_remove: 50
rollback: 50
| 1,173 | 28.35 | 78 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/cache.yaml | overrides:
ceph:
log-ignorelist:
- must scrub before tier agent can activate
conf:
osd:
# override short_pg_log_entries.yaml (which sets these under [global])
osd_min_pg_log_entries: 3000
osd_max_pg_log_entries: 3000
tasks:
- exec:
client.0:
- sudo ceph osd pool create base 4
- sudo ceph osd pool application enable base rados
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add base cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay base cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 3600
- sudo ceph osd pool set cache min_read_recency_for_promote 0
- sudo ceph osd pool set cache min_write_recency_for_promote 0
- rados:
clients: [client.0]
pools: [base]
ops: 4000
objects: 500
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
cache_flush: 50
cache_try_flush: 50
cache_evict: 50
| 1,119 | 29.27027 | 78 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/dedup-io-mixed.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create low_tier 4
- rados:
clients: [client.0]
low_tier_pool: 'low_tier'
ops: 1500
objects: 50
set_chunk: true
enable_dedup: true
dedup_chunk_size: '131072'
dedup_chunk_algo: 'fastcdc'
op_weights:
read: 100
write: 50
set_chunk: 30
tier_promote: 10
tier_flush: 5
tier_evict: 10
| 405 | 18.333333 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/dedup-io-snaps.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create low_tier 4
- rados:
clients: [client.0]
low_tier_pool: 'low_tier'
ops: 1500
objects: 50
set_chunk: true
enable_dedup: true
dedup_chunk_size: '131072'
dedup_chunk_algo: 'fastcdc'
op_weights:
read: 100
write: 50
set_chunk: 30
tier_promote: 10
tier_flush: 5
tier_evict: 10
snap_create: 10
snap_remove: 10
rollback: 10
| 468 | 18.541667 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml | overrides:
conf:
osd:
osd deep scrub update digest min age: 0
tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
pool_snaps: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 314 | 15.578947 | 45 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- \(POOL_APP_NOT_ENABLED\)
- \(PG_AVAILABILITY\)
crush_tunables: jewel
conf:
client:
debug ms: 1
debug objecter: 20
debug rados: 20
mon:
mon warn on pool no app: false
debug mgrc: 20
osd:
osd class load list: "*"
osd class default list: "*"
tasks:
- workunit:
clients:
client.0:
- rados/test.sh
| 468 | 18.541667 | 38 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/radosbench-high-concurrency.yaml | overrides:
ceph:
conf:
client.0:
debug ms: 1
debug objecter: 20
debug rados: 20
tasks:
- full_sequential:
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
- radosbench:
clients: [client.0]
concurrency: 128
size: 8192
time: 90
| 918 | 17.38 | 26 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/radosbench.yaml | overrides:
ceph:
conf:
client.0:
debug ms: 1
debug objecter: 20
debug rados: 20
tasks:
- full_sequential:
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
- radosbench:
clients: [client.0]
time: 90
| 427 | 16.12 | 26 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/redirect.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create low_tier 4
- rados:
clients: [client.0]
low_tier_pool: 'low_tier'
ops: 4000
objects: 500
set_redirect: true
op_weights:
read: 100
write: 100
delete: 50
copy_from: 50
| 277 | 16.375 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/redirect_promote_tests.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create low_tier 4
- rados:
clients: [client.0]
low_tier_pool: 'low_tier'
ops: 4000
objects: 500
set_redirect: true
op_weights:
set_redirect: 100
read: 50
tier_promote: 30
| 269 | 17 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/redirect_set_object.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create low_tier 4
- rados:
clients: [client.0]
low_tier_pool: 'low_tier'
ops: 4000
objects: 500
set_redirect: true
op_weights:
set_redirect: 100
copy_from: 100
| 252 | 17.071429 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/set-chunks-read.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create low_tier 4
- rados:
clients: [client.0]
low_tier_pool: 'low_tier'
ops: 4000
objects: 300
set_chunk: true
op_weights:
chunk_read: 100
tier_promote: 10
| 249 | 16.857143 | 44 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/small-objects-balanced.yaml | overrides:
ceph:
crush_tunables: jewel
tasks:
- rados:
clients: [client.0]
ops: 400000
max_seconds: 600
max_in_flight: 64
objects: 1024
size: 16384
balance_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
setattr: 25
rmattr: 25
| 386 | 15.826087 | 25 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/small-objects-localized.yaml | overrides:
ceph:
crush_tunables: jewel
tasks:
- rados:
clients: [client.0]
ops: 400000
max_seconds: 600
max_in_flight: 64
objects: 1024
size: 16384
localize_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
setattr: 25
rmattr: 25
| 387 | 15.869565 | 25 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/small-objects.yaml | overrides:
ceph:
crush_tunables: jewel
tasks:
- rados:
clients: [client.0]
ops: 400000
max_seconds: 600
max_in_flight: 64
objects: 1024
size: 16384
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
setattr: 25
rmattr: 25
| 362 | 15.5 | 25 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/snaps-few-objects-balanced.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
balance_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 243 | 15.266667 | 23 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/snaps-few-objects-localized.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
localize_reads: true
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 244 | 15.333333 | 24 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 50
op_weights:
read: 100
write: 100
delete: 50
snap_create: 50
snap_remove: 50
rollback: 50
copy_from: 50
| 219 | 14.714286 | 23 | yaml |
null | ceph-main/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml | tasks:
- rados:
clients: [client.0]
ops: 4000
objects: 500
write_fadvise_dontneed: true
op_weights:
write: 100
| 137 | 14.333333 | 32 | yaml |
null | ceph-main/qa/suites/rados/valgrind-leaks/1-start.yaml | openstack:
- volumes: # attached to each instance
count: 2
size: 10 # GB
overrides:
install:
ceph:
debuginfo: true
ceph:
log-ignorelist:
- overall HEALTH_
- \(PG_
conf:
global:
osd heartbeat grace: 40
osd max object name len: 460
osd max object namespace len: 64
mon:
mon osd crush smoke test: false
osd:
osd fast shutdown: false
valgrind:
mon: [--tool=memcheck, --leak-check=full, --show-reachable=yes]
osd: [--tool=memcheck]
roles:
- [mon.a, mon.b, mon.c, mgr.x, mgr.y, osd.0, osd.1, osd.2, client.0]
tasks:
- install:
- ceph:
| 652 | 20.064516 | 69 | yaml |
null | ceph-main/qa/suites/rados/valgrind-leaks/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/rados/valgrind-leaks/2-inject-leak/mon.yaml | overrides:
ceph:
expect_valgrind_errors: true
tasks:
- exec:
mon.a:
- ceph tell mon.a leak_some_memory
| 119 | 14 | 40 | yaml |
null | ceph-main/qa/suites/rados/valgrind-leaks/2-inject-leak/none.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/valgrind-leaks/2-inject-leak/osd.yaml | overrides:
ceph:
expect_valgrind_errors: true
tasks:
- exec:
mon.a:
- ceph tell osd.0 leak_some_memory
| 119 | 14 | 40 | yaml |
null | ceph-main/qa/suites/rados/verify/centos_latest.yaml | .qa/distros/supported/centos_latest.yaml | 40 | 40 | 40 | yaml |
null | ceph-main/qa/suites/rados/verify/ceph.yaml | overrides:
ceph:
conf:
mon:
mon min osdmap epochs: 50
paxos service trim min: 10
# prune full osdmaps regularly
mon osdmap full prune min: 15
mon osdmap full prune interval: 2
mon osdmap full prune txsize: 2
osd:
debug monc: 20
tasks:
- install:
- ceph:
| 328 | 19.5625 | 41 | yaml |
null | ceph-main/qa/suites/rados/verify/rados.yaml | .qa/config/rados.yaml | 21 | 21 | 21 | yaml |
null | ceph-main/qa/suites/rados/verify/clusters/fixed-2.yaml | .qa/clusters/fixed-2.yaml | 25 | 25 | 25 | yaml |
null | ceph-main/qa/suites/rados/verify/clusters/openstack.yaml | openstack:
- volumes: # attached to each instance
count: 4
size: 10 # GB
| 87 | 16.6 | 40 | yaml |
null | ceph-main/qa/suites/rados/verify/d-thrash/none.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rados/verify/d-thrash/default/default.yaml | overrides:
ceph:
log-ignorelist:
- but it is still running
- objects unfound and apparently lost
tasks:
- thrashosds:
timeout: 1200
chance_pgnum_grow: 1
chance_pgnum_shrink: 1
chance_pgpnum_fix: 1
| 227 | 18 | 41 | yaml |
null | ceph-main/qa/suites/rados/verify/d-thrash/default/thrashosds-health.yaml | .qa/tasks/thrashosds-health.yaml | 32 | 32 | 32 | yaml |
null | ceph-main/qa/suites/rados/verify/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rados/verify/tasks/mon_recovery.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(MON_DOWN\)
- \(OSDMAP_FLAGS\)
- \(SMALLER_PGP_NUM\)
- \(POOL_APP_NOT_ENABLED\)
- \(SLOW OPS\)
- slow request
tasks:
- mon_recovery:
| 235 | 17.153846 | 32 | yaml |
null | ceph-main/qa/suites/rados/verify/tasks/rados_api_tests.yaml | overrides:
ceph:
log-ignorelist:
- reached quota
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
- \(POOL_FULL\)
- \(SMALLER_PGP_NUM\)
- \(SLOW_OPS\)
- \(CACHE_POOL_NEAR_FULL\)
- \(POOL_APP_NOT_ENABLED\)
- \(PG_AVAILABILITY\)
- \(OBJECT_MISPLACED\)
- slow request
conf:
client:
debug ms: 1
debug objecter: 20
debug rados: 20
debug monc: 20
mon:
mon warn on pool no app: false
osd:
osd class load list: "*"
osd class default list: "*"
osd client watch timeout: 120
tasks:
- workunit:
timeout: 6h
env:
ALLOW_TIMEOUTS: "1"
clients:
client.0:
- rados/test.sh
| 743 | 20.257143 | 38 | yaml |
null | ceph-main/qa/suites/rados/verify/tasks/rados_cls_all.yaml | overrides:
ceph:
conf:
osd:
osd_class_load_list: "*"
osd_class_default_list: "*"
tasks:
- workunit:
clients:
client.0:
- cls
| 171 | 13.333333 | 35 | yaml |
null | ceph-main/qa/suites/rados/verify/validater/lockdep.yaml | overrides:
ceph:
conf:
global:
lockdep: true
| 65 | 10 | 21 | yaml |
null | ceph-main/qa/suites/rados/verify/validater/valgrind.yaml | # see http://tracker.ceph.com/issues/20360 and http://tracker.ceph.com/issues/18126
os_type: centos
overrides:
install:
ceph:
debuginfo: true
ceph:
conf:
global:
osd heartbeat grace: 80
mon:
mon osd crush smoke test: false
osd:
osd fast shutdown: false
debug bluestore: 1
debug bluefs: 1
log-ignorelist:
- overall HEALTH_
# valgrind is slow.. we might get PGs stuck peering etc
- \(PG_
# mons sometimes are left off of initial quorum due to valgrind slowness. ok to ignore here because we'll still catch an actual crash due to the core
- \(MON_DOWN\)
- \(SLOW_OPS\)
- slow request
valgrind:
mon: [--tool=memcheck, --leak-check=full, --show-reachable=yes]
osd: [--tool=memcheck]
mds: [--tool=memcheck]
# https://tracker.ceph.com/issues/38621
# mgr: [--tool=memcheck]
| 905 | 27.3125 | 150 | yaml |
null | ceph-main/qa/suites/rbd/basic/base/install.yaml | tasks:
- install:
- ceph:
| 26 | 5.75 | 10 | yaml |
null | ceph-main/qa/suites/rbd/basic/cachepool/none.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rbd/basic/cachepool/small.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NEAR_FULL\)
- \(CACHE_POOL_NO_HIT_SET\)
tasks:
- exec:
client.0:
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add rbd cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay rbd cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 60
- sudo ceph osd pool set cache target_max_objects 250
| 569 | 30.666667 | 59 | yaml |
null | ceph-main/qa/suites/rbd/basic/clusters/fixed-1.yaml | .qa/clusters/fixed-1.yaml | 25 | 25 | 25 | yaml |
null | ceph-main/qa/suites/rbd/basic/clusters/openstack.yaml | openstack:
- volumes: # attached to each instance
count: 3
size: 30 # GB
| 87 | 16.6 | 40 | yaml |
null | ceph-main/qa/suites/rbd/basic/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rbd/basic/tasks/rbd_api_tests_old_format.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
- \(POOL_APP_NOT_ENABLED\)
- is full \(reached quota
- \(POOL_FULL\)
tasks:
- workunit:
clients:
client.0:
- rbd/test_librbd.sh
| 261 | 17.714286 | 33 | yaml |
null | ceph-main/qa/suites/rbd/basic/tasks/rbd_cls_tests.yaml | tasks:
- workunit:
clients:
client.0:
- cls/test_cls_rbd.sh
- cls/test_cls_lock.sh
- cls/test_cls_journal.sh
| 143 | 17 | 33 | yaml |
null | ceph-main/qa/suites/rbd/basic/tasks/rbd_lock_and_fence.yaml | tasks:
- workunit:
clients:
client.0:
- rbd/test_lock_fence.sh
| 81 | 12.666667 | 32 | yaml |
null | ceph-main/qa/suites/rbd/basic/tasks/rbd_python_api_tests_old_format.yaml | overrides:
ceph:
log-ignorelist:
- \(SLOW_OPS\)
- slow request
tasks:
- workunit:
clients:
client.0:
- rbd/test_librbd_python.sh
| 165 | 14.090909 | 35 | yaml |
null | ceph-main/qa/suites/rbd/cli/base/install.yaml | tasks:
- install:
- ceph:
| 26 | 5.75 | 10 | yaml |
null | ceph-main/qa/suites/rbd/cli/features/defaults.yaml | overrides:
ceph:
conf:
client:
rbd default features: 61
| 76 | 11.833333 | 32 | yaml |
null | ceph-main/qa/suites/rbd/cli/features/journaling.yaml | overrides:
ceph:
conf:
client:
rbd default features: 125
| 77 | 12 | 33 | yaml |
null | ceph-main/qa/suites/rbd/cli/features/layering.yaml | overrides:
ceph:
conf:
client:
rbd default features: 1
| 75 | 11.666667 | 31 | yaml |
null | ceph-main/qa/suites/rbd/cli/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rbd/cli/pool/ec-data-pool.yaml | tasks:
- exec:
client.0:
- sudo ceph osd erasure-code-profile set teuthologyprofile crush-failure-domain=osd m=1 k=2
- sudo ceph osd pool create datapool 4 4 erasure teuthologyprofile
- sudo ceph osd pool set datapool allow_ec_overwrites true
- rbd pool init datapool
overrides:
thrashosds:
bdev_inject_crash: 2
bdev_inject_crash_probability: .5
ceph:
fs: xfs
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NO_HIT_SET\)
conf:
client:
rbd default data pool: datapool
osd: # force bluestore since it's required for ec overwrites
osd objectstore: bluestore
bluestore block size: 96636764160
enable experimental unrecoverable data corrupting features: "*"
osd debug randomize hobject sort order: false
# this doesn't work with failures bc the log writes are not atomic across the two backends
# bluestore bluefs env mirror: true
| 951 | 33 | 97 | yaml |
null | ceph-main/qa/suites/rbd/cli/pool/none.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rbd/cli/pool/replicated-data-pool.yaml | tasks:
- exec:
client.0:
- sudo ceph osd pool create datapool 4
- rbd pool init datapool
overrides:
ceph:
conf:
client:
rbd default data pool: datapool
| 189 | 14.833333 | 44 | yaml |
null | ceph-main/qa/suites/rbd/cli/pool/small-cache-pool.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NEAR_FULL\)
- \(CACHE_POOL_NO_HIT_SET\)
tasks:
- exec:
client.0:
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add rbd cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay rbd cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 60
- sudo ceph osd pool set cache target_max_objects 250
| 569 | 30.666667 | 59 | yaml |
null | ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_generic.yaml | tasks:
- workunit:
clients:
client.0:
- rbd/cli_generic.sh
| 77 | 12 | 28 | yaml |
null | ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_groups.yaml | tasks:
- workunit:
clients:
client.0:
- rbd/rbd_groups.sh
| 76 | 11.833333 | 27 | yaml |
null | ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_import_export.yaml | tasks:
- workunit:
clients:
client.0:
- rbd/import_export.sh
| 79 | 12.333333 | 30 | yaml |
null | ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_luks_encryption.yaml | overrides:
install:
ceph:
extra_packages: [rbd-nbd]
tasks:
- workunit:
clients:
client.0:
- rbd/luks-encryption.sh
| 145 | 13.6 | 32 | yaml |
null | ceph-main/qa/suites/rbd/cli/workloads/rbd_cli_migration.yaml | tasks:
- workunit:
clients:
client.0:
- rbd/cli_migration.sh
| 79 | 12.333333 | 30 | yaml |
null | ceph-main/qa/suites/rbd/cli_v1/base/install.yaml | tasks:
- install:
- ceph:
| 26 | 5.75 | 10 | yaml |
null | ceph-main/qa/suites/rbd/cli_v1/features/format-1.yaml | overrides:
ceph:
conf:
client:
rbd default format: 1
| 73 | 11.333333 | 29 | yaml |
null | ceph-main/qa/suites/rbd/cli_v1/msgr-failures/few.yaml | overrides:
ceph:
conf:
global:
ms inject socket failures: 5000
mon client directed command retry: 5
log-ignorelist:
- \(OSD_SLOW_PING_TIME
| 177 | 18.777778 | 44 | yaml |
null | ceph-main/qa/suites/rbd/cli_v1/pool/none.yaml | 0 | 0 | 0 | yaml |
|
null | ceph-main/qa/suites/rbd/cli_v1/pool/small-cache-pool.yaml | overrides:
ceph:
log-ignorelist:
- overall HEALTH_
- \(CACHE_POOL_NEAR_FULL\)
- \(CACHE_POOL_NO_HIT_SET\)
tasks:
- exec:
client.0:
- sudo ceph osd pool create cache 4
- sudo ceph osd tier add rbd cache
- sudo ceph osd tier cache-mode cache writeback
- sudo ceph osd tier set-overlay rbd cache
- sudo ceph osd pool set cache hit_set_type bloom
- sudo ceph osd pool set cache hit_set_count 8
- sudo ceph osd pool set cache hit_set_period 60
- sudo ceph osd pool set cache target_max_objects 250
| 569 | 30.666667 | 59 | yaml |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.