2026-04-01T13:49:06.849 INFO:root:teuthology version: 1.2.4.dev37+ga59626679 2026-04-01T13:49:06.854 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-04-01T13:49:06.878 INFO:teuthology.run:Config: archive_path: /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802 branch: wip-sse-s3-on-v20.2.0 description: rgw/dedup/{beast bluestore-bitmap fixed-3-rgw ignore-pg-availability overrides supported-distros/{rocky_latest} tasks/{0-install test_dedup}} email: null first_in_suite: false flavor: default job_id: '4802' last_in_suite: false machine_type: vps name: supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: rocky os_version: '9.7' overrides: admin_socket: branch: wip-sse-s3-on-v20.2.0 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: logical_volumes: lv_1: scratch_dev: true size: 25%VG vg: vg_nvme lv_2: scratch_dev: true size: 25%VG vg: vg_nvme lv_3: scratch_dev: true size: 25%VG vg: vg_nvme lv_4: scratch_dev: true size: 25%VG vg: vg_nvme timezone: UTC volume_groups: vg_nvme: pvs: /dev/vdb,/dev/vdc,/dev/vdd,/dev/vde ceph: conf: client: debug rgw: 20 debug rgw dedup: 20 setgroup: ceph setuser: ceph global: osd_max_pg_log_entries: 10 osd_min_pg_log_entries: 10 mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: bdev async discard: true bdev enable discard: true bluestore allocator: bitmap bluestore block size: 96636764160 bluestore fsck on mount: true debug bluefs: 1/20 debug bluestore: 1/20 debug ms: 1 debug osd: 20 debug rocksdb: 4/10 mon osd backfillfull_ratio: 0.85 mon osd full ratio: 0.9 mon osd nearfull ratio: 0.8 osd failsafe full ratio: 0.95 osd mclock iops capacity threshold hdd: 49000 osd objectstore: bluestore osd shutdown pgref assert: true flavor: default fs: xfs log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - \(PG_AVAILABILITY\) - \(PG_DEGRADED\) - \(POOL_APP_NOT_ENABLED\) - not have an application enabled sha1: e2ce8426bc5387b913d1b06bd31280ced108bbb6 ceph-deploy: bluestore: true conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} osd: bdev async discard: true bdev enable discard: true bluestore block size: 96636764160 bluestore fsck on mount: true debug bluefs: 1/20 debug bluestore: 1/20 debug rocksdb: 4/10 mon osd backfillfull_ratio: 0.85 mon osd full ratio: 0.9 mon osd nearfull ratio: 0.8 osd failsafe full ratio: 0.95 osd objectstore: bluestore fs: xfs cephadm: cephadm_binary_url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:sse-s3-kmip-preview-not-for-production-2 install: ceph: flavor: default sha1: e2ce8426bc5387b913d1b06bd31280ced108bbb6 extra_system_packages: deb: - python3-jmespath - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-jmespath - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64 rgw: frontend: beast storage classes: FROZEN: null LUKEWARM: null s3tests: sha1: e0c4ff71baef6d5126a0201df5fe54196d89b296 selinux: allowlist: - scontext=system_u:system_r:getty_t:s0 thrashosds: bdev_inject_crash: 2 bdev_inject_crash_probability: 0.5 workunit: branch: tt-20.2.0-sse-s3-kmip-preview-not-for-production-2 sha1: 0d46c98b681455ef9cd34c2d3c6981da4e9f1aea owner: supriti priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - mon.a - mon.c - mgr.y - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - - mon.b - mgr.x - osd.4 - osd.5 - osd.6 - osd.7 - client.1 - - client.2 seed: 343 sha1: e2ce8426bc5387b913d1b06bd31280ced108bbb6 sleep_before_teardown: 0 suite: rgw suite_branch: tt-20.2.0-sse-s3-kmip-preview-not-for-production-2 suite_path: /home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa suite_relpath: qa suite_repo: http://git.local/ceph.git suite_sha1: 0d46c98b681455ef9cd34c2d3c6981da4e9f1aea targets: vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNTOLJDVW6KK/6J5BdWj48jn2txkC0z7tw+YqaQ/6nXMOyhu/56c5vRw26fnzDUT+rhL+2WOGbcS6a3VQB//jnI= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAR1cnfhNFk7xCg/QzMr9TAXWR2qIkMIewcDvjVdOdAwsgIREaYdKXhUYOz+HPynWcoyzk33kjDDD7Q8PRbkzjE= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPVr7OG9quN7TJQPoyWRtARk297W6aw+/g+hvz++NgJFauXwO5rB98MdTAsBBRadXZsdskUFGesjgEbhYM319m0= tasks: - install: null - ceph: null - openssl_keys: null - rgw: - client.0 - client.1 - client.2 - tox: - client.0 - tox: - client.0 - dedup-tests: client.0: rgw_server: client.0 teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: uv2 teuthology_repo: https://github.com/kshtsk/teuthology teuthology_sha1: a59626679648f962bca99d20d35578f2998c8f37 timestamp: 2026-04-01_13:45:16 tube: vps user: supriti verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.1399776 2026-04-01T13:49:06.878 INFO:teuthology.run:suite_path is set to /home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa; will attempt to use it 2026-04-01T13:49:06.878 INFO:teuthology.run:Found tasks at /home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks 2026-04-01T13:49:06.878 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-04-01T13:49:06.879 INFO:teuthology.task.internal:Saving configuration 2026-04-01T13:49:06.908 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-04-01T13:49:06.909 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-04-01T13:49:06.915 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'rocky', 'os_version': '9.7', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-04-01 13:47:38.218321', 'locked_by': 'supriti', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNTOLJDVW6KK/6J5BdWj48jn2txkC0z7tw+YqaQ/6nXMOyhu/56c5vRw26fnzDUT+rhL+2WOGbcS6a3VQB//jnI='} 2026-04-01T13:49:06.922 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'rocky', 'os_version': '9.7', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-04-01 13:47:38.217468', 'locked_by': 'supriti', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAR1cnfhNFk7xCg/QzMr9TAXWR2qIkMIewcDvjVdOdAwsgIREaYdKXhUYOz+HPynWcoyzk33kjDDD7Q8PRbkzjE='} 2026-04-01T13:49:06.927 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'rocky', 'os_version': '9.7', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-04-01 13:47:38.218000', 'locked_by': 'supriti', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPVr7OG9quN7TJQPoyWRtARk297W6aw+/g+hvz++NgJFauXwO5rB98MdTAsBBRadXZsdskUFGesjgEbhYM319m0='} 2026-04-01T13:49:06.927 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-04-01T13:49:06.928 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['mon.a', 'mon.c', 'mgr.y', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'] 2026-04-01T13:49:06.928 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['mon.b', 'mgr.x', 'osd.4', 'osd.5', 'osd.6', 'osd.7', 'client.1'] 2026-04-01T13:49:06.928 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['client.2'] 2026-04-01T13:49:06.928 INFO:teuthology.run_tasks:Running task console_log... 2026-04-01T13:49:06.936 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-04-01T13:49:06.943 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-04-01T13:49:06.948 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-04-01T13:49:06.948 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fd8b0c61240>, signals=[15]) 2026-04-01T13:49:06.948 INFO:teuthology.run_tasks:Running task internal.connect... 2026-04-01T13:49:06.949 INFO:teuthology.task.internal:Opening connections... 2026-04-01T13:49:06.949 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-04-01T13:49:06.950 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-01T13:49:07.009 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-04-01T13:49:07.009 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-01T13:49:07.070 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-04-01T13:49:07.071 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-01T13:49:07.131 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-04-01T13:49:07.132 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-04-01T13:49:07.146 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-04-01T13:49:07.146 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:NAME="Rocky Linux" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9.7 (Blue Onyx)" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:ID="rocky" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel centos fedora" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9.7" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="Rocky Linux 9.7 (Blue Onyx)" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;32" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:rocky:rocky:9::baseos" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://rockylinux.org/" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:VENDOR_NAME="RESF" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:VENDOR_URL="https://resf.org/" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://bugs.rockylinux.org/" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:SUPPORT_END="2032-05-31" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:ROCKY_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Rocky Linux" 2026-04-01T13:49:07.203 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-01T13:49:07.204 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-04-01T13:49:07.208 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-04-01T13:49:07.227 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-04-01T13:49:07.227 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:NAME="Rocky Linux" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9.7 (Blue Onyx)" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:ID="rocky" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel centos fedora" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9.7" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="Rocky Linux 9.7 (Blue Onyx)" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;32" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:rocky:rocky:9::baseos" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://rockylinux.org/" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:VENDOR_NAME="RESF" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:VENDOR_URL="https://resf.org/" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://bugs.rockylinux.org/" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:SUPPORT_END="2032-05-31" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:ROCKY_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Rocky Linux" 2026-04-01T13:49:07.285 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-01T13:49:07.285 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-04-01T13:49:07.291 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-04-01T13:49:07.305 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-04-01T13:49:07.306 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:NAME="Rocky Linux" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="9.7 (Blue Onyx)" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:ID="rocky" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE="rhel centos fedora" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="9.7" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:PLATFORM_ID="platform:el9" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="Rocky Linux 9.7 (Blue Onyx)" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:ANSI_COLOR="0;32" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:LOGO="fedora-logo-icon" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:CPE_NAME="cpe:/o:rocky:rocky:9::baseos" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://rockylinux.org/" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:VENDOR_NAME="RESF" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:VENDOR_URL="https://resf.org/" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://bugs.rockylinux.org/" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:SUPPORT_END="2032-05-31" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:ROCKY_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT="Rocky Linux" 2026-04-01T13:49:07.362 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-01T13:49:07.362 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-04-01T13:49:07.366 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-04-01T13:49:07.368 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-04-01T13:49:07.369 INFO:teuthology.task.internal:Checking for old test directory... 2026-04-01T13:49:07.369 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-04-01T13:49:07.371 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-04-01T13:49:07.372 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-04-01T13:49:07.417 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-04-01T13:49:07.418 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-04-01T13:49:07.418 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-04-01T13:49:07.428 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-04-01T13:49:07.430 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-04-01T13:49:07.442 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-04-01T13:49:07.446 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-04-01T13:49:07.472 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-04-01T13:49:07.473 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-04-01T13:49:07.481 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-04-01T13:49:07.495 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T13:49:07.686 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-04-01T13:49:07.703 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T13:49:07.903 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-04-01T13:49:07.917 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T13:49:08.106 INFO:teuthology.run_tasks:Running task internal.base... 2026-04-01T13:49:08.107 INFO:teuthology.task.internal:Creating test directory... 2026-04-01T13:49:08.107 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-04-01T13:49:08.109 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-04-01T13:49:08.110 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-04-01T13:49:08.128 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-04-01T13:49:08.129 INFO:teuthology.run_tasks:Running task internal.archive... 2026-04-01T13:49:08.130 INFO:teuthology.task.internal:Creating archive directory... 2026-04-01T13:49:08.130 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-04-01T13:49:08.167 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-04-01T13:49:08.169 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-04-01T13:49:08.191 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-04-01T13:49:08.192 INFO:teuthology.task.internal:Enabling coredump saving... 2026-04-01T13:49:08.192 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-04-01T13:49:08.240 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T13:49:08.240 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-04-01T13:49:08.257 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T13:49:08.257 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-04-01T13:49:08.273 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T13:49:08.273 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-04-01T13:49:08.281 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-04-01T13:49:08.299 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-04-01T13:49:08.306 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-01T13:49:08.316 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-01T13:49:08.323 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-01T13:49:08.333 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-01T13:49:08.338 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-01T13:49:08.346 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-01T13:49:08.347 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-04-01T13:49:08.348 INFO:teuthology.task.internal:Configuring sudo... 2026-04-01T13:49:08.348 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-04-01T13:49:08.359 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-04-01T13:49:08.377 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-04-01T13:49:08.411 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-04-01T13:49:08.413 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-04-01T13:49:08.413 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-04-01T13:49:08.426 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-04-01T13:49:08.444 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-04-01T13:49:08.469 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-04-01T13:49:08.506 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-04-01T13:49:08.565 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:49:08.565 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-04-01T13:49:08.624 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-04-01T13:49:08.651 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-04-01T13:49:08.708 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:49:08.708 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-04-01T13:49:08.766 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-04-01T13:49:08.797 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-04-01T13:49:08.856 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:49:08.856 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-04-01T13:49:08.922 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-04-01T13:49:08.924 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-04-01T13:49:08.926 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-04-01T13:49:08.953 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-01T13:49:08.955 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-01T13:49:08.995 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-01T13:49:09.481 INFO:teuthology.run_tasks:Running task internal.timer... 2026-04-01T13:49:09.483 INFO:teuthology.task.internal:Starting timer... 2026-04-01T13:49:09.483 INFO:teuthology.run_tasks:Running task pcp... 2026-04-01T13:49:09.486 INFO:teuthology.run_tasks:Running task selinux... 2026-04-01T13:49:09.488 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:getty_t:s0']} 2026-04-01T13:49:09.488 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-04-01T13:49:09.488 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-04-01T13:49:09.488 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-04-01T13:49:09.488 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-04-01T13:49:09.488 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-04-01T13:49:09.488 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-04-01T13:49:09.488 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-04-01T13:49:09.489 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}} 2026-04-01T13:49:09.489 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-04-01T13:49:09.491 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-04-01T13:49:10.080 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-04-01T13:49:10.087 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-04-01T13:49:10.088 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "logical_volumes": {"lv_1": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_2": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_3": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_4": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}}, "timezone": "UTC", "volume_groups": {"vg_nvme": {"pvs": "/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde"}}}' -i /tmp/teuth_ansible_inventorym4_gtmtd --limit vm06.local,vm08.local,vm09.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-04-01T13:53:06.930 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm08.local'), Remote(name='ubuntu@vm09.local')] 2026-04-01T13:53:06.930 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-04-01T13:53:06.931 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-01T13:53:06.993 DEBUG:teuthology.orchestra.run.vm06:> true 2026-04-01T13:53:07.076 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-04-01T13:53:07.076 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-04-01T13:53:07.077 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-01T13:53:07.140 DEBUG:teuthology.orchestra.run.vm08:> true 2026-04-01T13:53:07.233 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-04-01T13:53:07.233 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-04-01T13:53:07.234 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-01T13:53:07.295 DEBUG:teuthology.orchestra.run.vm09:> true 2026-04-01T13:53:07.379 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-04-01T13:53:07.380 INFO:teuthology.run_tasks:Running task clock... 2026-04-01T13:53:07.382 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-04-01T13:53:07.382 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-04-01T13:53:07.382 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-01T13:53:07.384 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-04-01T13:53:07.384 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-01T13:53:07.386 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-04-01T13:53:07.386 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-01T13:53:07.413 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-04-01T13:53:07.416 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-04-01T13:53:07.429 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-04-01T13:53:07.431 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-04-01T13:53:07.449 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-04-01T13:53:07.457 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-04-01T13:53:07.458 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-04-01T13:53:07.463 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-04-01T13:53:07.470 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-04-01T13:53:07.472 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-04-01T13:53:07.484 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-04-01T13:53:07.487 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-04-01T13:53:07.487 INFO:teuthology.orchestra.run.vm09.stderr:sudo: ntpd: command not found 2026-04-01T13:53:07.498 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-04-01T13:53:07.500 INFO:teuthology.orchestra.run.vm09.stdout:506 Cannot talk to daemon 2026-04-01T13:53:07.503 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-04-01T13:53:07.514 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-04-01T13:53:07.528 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-04-01T13:53:07.549 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-04-01T13:53:07.552 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-01T13:53:07.552 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-04-01T13:53:07.552 INFO:teuthology.orchestra.run.vm08.stdout:^? node-3.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.552 INFO:teuthology.orchestra.run.vm08.stdout:^? ntp.b-ite.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.552 INFO:teuthology.orchestra.run.vm08.stdout:^? 185.13.148.71 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.552 INFO:teuthology.orchestra.run.vm08.stdout:^? byggvir.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.559 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-04-01T13:53:07.561 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-01T13:53:07.561 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-04-01T13:53:07.561 INFO:teuthology.orchestra.run.vm06.stdout:^? ntp.b-ite.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.561 INFO:teuthology.orchestra.run.vm06.stdout:^? 185.13.148.71 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.561 INFO:teuthology.orchestra.run.vm06.stdout:^? byggvir.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.561 INFO:teuthology.orchestra.run.vm06.stdout:^? node-3.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.583 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-04-01T13:53:07.587 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-01T13:53:07.587 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-04-01T13:53:07.587 INFO:teuthology.orchestra.run.vm09.stdout:^? ntp.b-ite.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.587 INFO:teuthology.orchestra.run.vm09.stdout:^? 185.13.148.71 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.587 INFO:teuthology.orchestra.run.vm09.stdout:^? byggvir.de 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.587 INFO:teuthology.orchestra.run.vm09.stdout:^? node-3.infogral.is 0 6 0 - +0ns[ +0ns] +/- 0ns 2026-04-01T13:53:07.588 INFO:teuthology.run_tasks:Running task install... 2026-04-01T13:53:07.589 DEBUG:teuthology.task.install:project ceph 2026-04-01T13:53:07.589 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}]} 2026-04-01T13:53:07.590 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6', 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-04-01T13:53:07.590 INFO:teuthology.task.install:Using flavor: default 2026-04-01T13:53:07.592 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-04-01T13:53:07.592 INFO:teuthology.task.install:extra packages: [] 2026-04-01T13:53:07.592 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}]} 2026-04-01T13:53:07.592 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}] 2026-04-01T13:53:07.592 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:53:07.592 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-04-01T13:53:07.592 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}]} 2026-04-01T13:53:07.592 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}] 2026-04-01T13:53:07.592 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:53:07.592 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-04-01T13:53:07.593 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}]} 2026-04-01T13:53:07.593 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}] 2026-04-01T13:53:07.593 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:53:07.593 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-04-01T13:53:07.622 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:53:07.622 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-04-01T13:53:07.639 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:53:07.639 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-04-01T13:53:07.658 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:53:07.658 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-04-01T13:53:07.690 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:53:07.690 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-04-01T13:53:07.710 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:53:07.710 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-04-01T13:53:07.728 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:53:07.728 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-04-01T13:53:07.758 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-jmespath, python3-xmltodict, s3cmd on remote rpm x86_64 2026-04-01T13:53:07.759 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-04-01T13:53:07.781 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-jmespath, python3-xmltodict, s3cmd on remote rpm x86_64 2026-04-01T13:53:07.781 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-04-01T13:53:07.799 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-jmespath, python3-xmltodict, s3cmd on remote rpm x86_64 2026-04-01T13:53:07.799 DEBUG:teuthology.orchestra.run.vm09:> sudo yum clean all 2026-04-01T13:53:07.940 INFO:teuthology.orchestra.run.vm08.stdout:47 files removed 2026-04-01T13:53:07.959 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-jmespath python3-xmltodict s3cmd 2026-04-01T13:53:07.975 INFO:teuthology.orchestra.run.vm09.stdout:47 files removed 2026-04-01T13:53:07.986 INFO:teuthology.orchestra.run.vm06.stdout:47 files removed 2026-04-01T13:53:07.999 DEBUG:teuthology.orchestra.run.vm09:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-jmespath python3-xmltodict s3cmd 2026-04-01T13:53:08.013 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-jmespath python3-xmltodict s3cmd 2026-04-01T13:53:08.695 INFO:teuthology.orchestra.run.vm08.stdout:ceph 163 kB/s | 90 kB 00:00 2026-04-01T13:53:08.702 INFO:teuthology.orchestra.run.vm06.stdout:ceph 172 kB/s | 90 kB 00:00 2026-04-01T13:53:08.730 INFO:teuthology.orchestra.run.vm09.stdout:ceph 165 kB/s | 90 kB 00:00 2026-04-01T13:53:08.957 INFO:teuthology.orchestra.run.vm08.stdout:ceph-noarch 107 kB/s | 25 kB 00:00 2026-04-01T13:53:08.963 INFO:teuthology.orchestra.run.vm06.stdout:ceph-noarch 106 kB/s | 25 kB 00:00 2026-04-01T13:53:09.186 INFO:teuthology.orchestra.run.vm08.stdout:ceph-source 11 kB/s | 2.3 kB 00:00 2026-04-01T13:53:09.271 INFO:teuthology.orchestra.run.vm09.stdout:ceph-noarch 48 kB/s | 25 kB 00:00 2026-04-01T13:53:09.888 INFO:teuthology.orchestra.run.vm09.stdout:ceph-source 3.8 kB/s | 2.3 kB 00:00 2026-04-01T13:53:09.956 INFO:teuthology.orchestra.run.vm08.stdout:Extra Packages for Enterprise Linux 27 MB/s | 20 MB 00:00 2026-04-01T13:53:10.291 INFO:teuthology.orchestra.run.vm06.stdout:ceph-source 1.8 kB/s | 2.3 kB 00:01 2026-04-01T13:53:10.395 INFO:teuthology.orchestra.run.vm09.stdout:Extra Packages for Enterprise Linux 42 MB/s | 20 MB 00:00 2026-04-01T13:53:11.277 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 21 MB/s | 20 MB 00:00 2026-04-01T13:53:14.645 INFO:teuthology.orchestra.run.vm08.stdout:lab-extras 57 kB/s | 50 kB 00:00 2026-04-01T13:53:15.040 INFO:teuthology.orchestra.run.vm09.stdout:lab-extras 57 kB/s | 50 kB 00:00 2026-04-01T13:53:15.550 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - BaseOS 21 MB/s | 17 MB 00:00 2026-04-01T13:53:15.966 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 57 kB/s | 50 kB 00:00 2026-04-01T13:53:15.994 INFO:teuthology.orchestra.run.vm09.stdout:Rocky Linux 9 - BaseOS 19 MB/s | 17 MB 00:00 2026-04-01T13:53:16.824 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - BaseOS 22 MB/s | 17 MB 00:00 2026-04-01T13:53:17.428 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - AppStream 24 MB/s | 17 MB 00:00 2026-04-01T13:53:17.806 INFO:teuthology.orchestra.run.vm09.stdout:Rocky Linux 9 - AppStream 23 MB/s | 17 MB 00:00 2026-04-01T13:53:18.654 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - AppStream 23 MB/s | 17 MB 00:00 2026-04-01T13:53:19.706 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - CRB 7.9 MB/s | 4.3 MB 00:00 2026-04-01T13:53:20.041 INFO:teuthology.orchestra.run.vm09.stdout:Rocky Linux 9 - CRB 8.0 MB/s | 4.3 MB 00:00 2026-04-01T13:53:20.603 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - Extras 49 kB/s | 17 kB 00:00 2026-04-01T13:53:20.921 INFO:teuthology.orchestra.run.vm09.stdout:Rocky Linux 9 - Extras 50 kB/s | 17 kB 00:00 2026-04-01T13:53:20.982 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - CRB 7.6 MB/s | 4.3 MB 00:00 2026-04-01T13:53:21.727 INFO:teuthology.orchestra.run.vm08.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-01T13:53:21.727 INFO:teuthology.orchestra.run.vm08.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-01T13:53:21.750 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================= 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================= 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: bzip2 x86_64 1.0.8-10.el9_5 baseos 51 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 6.5 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 5.9 M 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 941 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 154 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 961 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 173 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 15 M 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 7.4 M 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 50 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 M 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 85 M 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 297 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 1.0 M 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 34 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 868 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 126 k 2026-04-01T13:53:21.755 INFO:teuthology.orchestra.run.vm08.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 267 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 163 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath noarch 1.0.1-1.el9_7 appstream 43 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 316 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 304 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 99 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 91 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.9 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 180 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout:Upgrading: 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 3.5 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.8 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9_7 appstream 104 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: c-ares x86_64 1.19.1-2.el9_4 baseos 110 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 43 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.3 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 289 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 5.0 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 17 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 17 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 25 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.7.2-4.el9 baseos 310 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 appstream 30 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 appstream 3.0 M 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 appstream 15 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: fuse x86_64 2.9.9-17.el9 baseos 78 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 41 k 2026-04-01T13:53:21.756 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-proxy2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 164 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 baseos 71 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-11.el9 baseos 794 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 appstream 171 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 159 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-11.el9 baseos 184 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 44 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 250 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 6.4 M 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 243 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-13.el9_6 appstream 239 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: lmdb-libs x86_64 0.9.29-3.el9 baseos 60 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 282 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 appstream 187 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 crb 21 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9.0.2 baseos 32 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 appstream 41 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 baseos 92 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: perl-Benchmark noarch 1.23-481.1.el9_6 appstream 25 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9_7 appstream 1.0 M 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 crb 862 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 appstream 5.8 M 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-04-01T13:53:21.757 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 45 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 163 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 241 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-5.el9 epel 173 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.10.0-5.el9 epel 290 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 baseos 1.2 M 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.23-2.el9 appstream 205 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 appstream 228 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 166 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 32 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 appstream 5.8 M 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 appstream 368 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 appstream 69 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9.0.1 baseos 103 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-04-01T13:53:21.758 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9_7 appstream 237 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 appstream 132 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 appstream 210 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 baseos 124 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 baseos 150 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9_6 baseos 115 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 43 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9.0.1 appstream 44 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 baseos 191 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 24.09.0-1.el9 appstream 221 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 65 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 appstream 299 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 baseos 180 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 188 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 89 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 63 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 baseos 263 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout:Installing weak dependencies: 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 22 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-daemon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 35 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli x86_64 2.13-1.el9 baseos 1.0 M 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-04-01T13:53:21.759 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 24.09.0-1.el9 appstream 36 k 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools x86_64 1:7.2-9.el9 baseos 551 k 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================= 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout:Install 150 Packages 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout:Upgrade 2 Packages 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 274 M 2026-04-01T13:53:21.760 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-04-01T13:53:21.865 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - Extras 50 kB/s | 17 kB 00:00 2026-04-01T13:53:22.045 INFO:teuthology.orchestra.run.vm09.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-01T13:53:22.046 INFO:teuthology.orchestra.run.vm09.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-01T13:53:22.069 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-04-01T13:53:22.073 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================================= 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================================= 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: bzip2 x86_64 1.0.8-10.el9_5 baseos 51 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 6.5 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 5.9 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 941 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 154 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 961 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 173 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 15 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 7.4 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 50 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 85 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 297 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 1.0 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 34 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 868 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 126 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 267 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 163 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath noarch 1.0.1-1.el9_7 appstream 43 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 316 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 304 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 99 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 91 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.9 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 180 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout:Upgrading: 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: librados2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 3.5 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: librbd1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.8 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options x86_64 1.75.0-13.el9_7 appstream 104 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: c-ares x86_64 1.19.1-2.el9_4 baseos 110 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 43 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.3 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 289 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 5.0 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 17 M 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 17 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 25 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup x86_64 2.7.2-4.el9 baseos 310 k 2026-04-01T13:53:22.074 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 appstream 30 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 appstream 3.0 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 appstream 15 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: fuse x86_64 2.9.9-17.el9 baseos 78 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 41 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-proxy2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 164 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libconfig x86_64 1.7.2-9.el9 baseos 71 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran x86_64 11.5.0-11.el9 baseos 794 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libnbd x86_64 1.20.3-4.el9 appstream 171 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 159 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath x86_64 11.5.0-11.el9 baseos 184 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 44 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 250 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: librgw2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 6.4 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 243 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: libxslt x86_64 1.1.34-13.el9_6 appstream 239 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: lmdb-libs x86_64 0.9.29-3.el9 baseos 60 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 282 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: lua x86_64 5.4.4-4.el9 appstream 187 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel x86_64 5.4.4-4.el9 crb 21 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: mailcap noarch 2.1.49-5.el9.0.2 baseos 32 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: openblas x86_64 0.3.29-1.el9 appstream 41 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: pciutils x86_64 3.7.0-7.el9 baseos 92 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: perl-Benchmark noarch 1.23-481.1.el9_6 appstream 25 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: protobuf x86_64 3.14.0-17.el9_7 appstream 1.0 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 crb 862 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel noarch 2.9.1-2.el9 appstream 5.8 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 45 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 163 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 241 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot noarch 10.0.1-5.el9 epel 173 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy noarch 18.10.0-5.el9 epel 290 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 baseos 1.2 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel x86_64 3.9.23-2.el9 appstream 205 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-04-01T13:53:22.075 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 appstream 228 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 166 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 32 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 appstream 5.8 M 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 appstream 368 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging noarch 20.9-5.el9 appstream 69 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply noarch 3.11-14.el9.0.1 baseos 103 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf noarch 3.14.0-17.el9_7 appstream 237 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 appstream 132 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 appstream 210 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser noarch 2.20-6.el9 baseos 124 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 baseos 150 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests noarch 2.25.1-10.el9_6 baseos 115 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 43 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml noarch 0.10.2-6.el9.0.1 appstream 44 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 baseos 191 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: qatlib x86_64 24.09.0-1.el9 appstream 221 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 65 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: socat x86_64 1.7.4.1-8.el9 appstream 299 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: unzip x86_64 6.0-59.el9 baseos 180 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 188 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 89 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 63 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: zip x86_64 3.0-35.el9 baseos 263 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout:Installing weak dependencies: 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 22 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-daemon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 35 k 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli x86_64 2.13-1.el9 baseos 1.0 M 2026-04-01T13:53:22.076 INFO:teuthology.orchestra.run.vm09.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service x86_64 24.09.0-1.el9 appstream 36 k 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools x86_64 1:7.2-9.el9 baseos 551 k 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout:============================================================================================= 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout:Install 150 Packages 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout:Upgrade 2 Packages 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 274 M 2026-04-01T13:53:22.077 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-04-01T13:53:22.744 INFO:teuthology.orchestra.run.vm09.stdout:(1/152): ceph-20.2.0-9.ge2ce8426bc5.el9.clyso.x 78 kB/s | 6.5 kB 00:00 2026-04-01T13:53:22.877 INFO:teuthology.orchestra.run.vm08.stdout:(1/152): ceph-20.2.0-9.ge2ce8426bc5.el9.clyso.x 66 kB/s | 6.5 kB 00:00 2026-04-01T13:53:22.986 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-01T13:53:22.986 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-01T13:53:23.008 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-04-01T13:53:23.012 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================= 2026-04-01T13:53:23.012 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-04-01T13:53:23.012 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================= 2026-04-01T13:53:23.012 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-04-01T13:53:23.012 INFO:teuthology.orchestra.run.vm06.stdout: bzip2 x86_64 1.0.8-10.el9_5 baseos 51 k 2026-04-01T13:53:23.012 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 6.5 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 5.9 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 941 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 154 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 961 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 173 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 15 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 7.4 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 50 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 85 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 297 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 1.0 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 34 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 868 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 126 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 267 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 163 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9_7 appstream 43 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 316 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 304 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 99 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 91 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.9 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 180 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 3.5 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.8 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9_7 appstream 104 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: c-ares x86_64 1.19.1-2.el9_4 baseos 110 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 43 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 2.3 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 289 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 5.0 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 17 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 17 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 25 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.7.2-4.el9 baseos 310 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 appstream 30 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 appstream 3.0 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 appstream 15 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: fuse x86_64 2.9.9-17.el9 baseos 78 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 41 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-proxy2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 24 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 164 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 71 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-11.el9 baseos 794 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 171 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 159 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-11.el9 baseos 184 k 2026-04-01T13:53:23.013 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 44 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 250 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 6.4 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 243 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-13.el9_6 appstream 239 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: lmdb-libs x86_64 0.9.29-3.el9 baseos 60 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 282 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 187 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 21 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9.0.2 baseos 32 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 41 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 92 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: perl-Benchmark noarch 1.23-481.1.el9_6 appstream 25 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9_7 appstream 1.0 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 crb 862 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 5.8 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 45 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 163 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 241 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-5.el9 epel 173 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.10.0-5.el9 epel 290 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 baseos 1.2 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.23-2.el9 appstream 205 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 appstream 228 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 166 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 32 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 appstream 5.8 M 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 appstream 368 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 69 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9.0.1 baseos 103 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9_7 appstream 237 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 appstream 132 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 appstream 210 k 2026-04-01T13:53:23.014 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 124 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 baseos 150 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9_6 baseos 115 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 43 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9.0.1 appstream 44 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 baseos 191 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 24.09.0-1.el9 appstream 221 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 65 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 299 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 180 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 188 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 89 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 63 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 263 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph-noarch 22 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso ceph 35 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.13-1.el9 baseos 1.0 M 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 24.09.0-1.el9 appstream 36 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools x86_64 1:7.2-9.el9 baseos 551 k 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================= 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:Install 150 Packages 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 274 M 2026-04-01T13:53:23.015 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-04-01T13:53:23.219 INFO:teuthology.orchestra.run.vm09.stdout:(2/152): ceph-base-20.2.0-9.ge2ce8426bc5.el9.cl 11 MB/s | 5.9 MB 00:00 2026-04-01T13:53:23.281 INFO:teuthology.orchestra.run.vm09.stdout:(3/152): ceph-immutable-object-cache-20.2.0-9.g 2.4 MB/s | 154 kB 00:00 2026-04-01T13:53:23.362 INFO:teuthology.orchestra.run.vm09.stdout:(4/152): ceph-fuse-20.2.0-9.ge2ce8426bc5.el9.cl 1.5 MB/s | 941 kB 00:00 2026-04-01T13:53:23.378 INFO:teuthology.orchestra.run.vm08.stdout:(2/152): ceph-fuse-20.2.0-9.ge2ce8426bc5.el9.cl 1.8 MB/s | 941 kB 00:00 2026-04-01T13:53:23.433 INFO:teuthology.orchestra.run.vm08.stdout:(3/152): ceph-immutable-object-cache-20.2.0-9.g 2.7 MB/s | 154 kB 00:00 2026-04-01T13:53:23.490 INFO:teuthology.orchestra.run.vm09.stdout:(5/152): ceph-mgr-20.2.0-9.ge2ce8426bc5.el9.cly 7.3 MB/s | 961 kB 00:00 2026-04-01T13:53:23.499 INFO:teuthology.orchestra.run.vm09.stdout:(6/152): ceph-mds-20.2.0-9.ge2ce8426bc5.el9.cly 11 MB/s | 2.3 MB 00:00 2026-04-01T13:53:23.623 INFO:teuthology.orchestra.run.vm08.stdout:(4/152): ceph-mds-20.2.0-9.ge2ce8426bc5.el9.cly 12 MB/s | 2.3 MB 00:00 2026-04-01T13:53:23.642 INFO:teuthology.orchestra.run.vm08.stdout:(5/152): ceph-base-20.2.0-9.ge2ce8426bc5.el9.cl 6.8 MB/s | 5.9 MB 00:00 2026-04-01T13:53:23.720 INFO:teuthology.orchestra.run.vm06.stdout:(1/152): ceph-20.2.0-9.ge2ce8426bc5.el9.clyso.x 64 kB/s | 6.5 kB 00:00 2026-04-01T13:53:23.720 INFO:teuthology.orchestra.run.vm08.stdout:(6/152): ceph-mgr-20.2.0-9.ge2ce8426bc5.el9.cly 9.7 MB/s | 961 kB 00:00 2026-04-01T13:53:23.829 INFO:teuthology.orchestra.run.vm09.stdout:(7/152): ceph-common-20.2.0-9.ge2ce8426bc5.el9. 20 MB/s | 24 MB 00:01 2026-04-01T13:53:24.056 INFO:teuthology.orchestra.run.vm08.stdout:(7/152): ceph-common-20.2.0-9.ge2ce8426bc5.el9. 19 MB/s | 24 MB 00:01 2026-04-01T13:53:24.096 INFO:teuthology.orchestra.run.vm09.stdout:(8/152): ceph-mon-20.2.0-9.ge2ce8426bc5.el9.cly 8.3 MB/s | 5.0 MB 00:00 2026-04-01T13:53:24.173 INFO:teuthology.orchestra.run.vm09.stdout:(9/152): ceph-selinux-20.2.0-9.ge2ce8426bc5.el9 328 kB/s | 25 kB 00:00 2026-04-01T13:53:24.239 INFO:teuthology.orchestra.run.vm06.stdout:(2/152): ceph-fuse-20.2.0-9.ge2ce8426bc5.el9.cl 1.8 MB/s | 941 kB 00:00 2026-04-01T13:53:24.328 INFO:teuthology.orchestra.run.vm06.stdout:(3/152): ceph-immutable-object-cache-20.2.0-9.g 1.7 MB/s | 154 kB 00:00 2026-04-01T13:53:24.536 INFO:teuthology.orchestra.run.vm06.stdout:(4/152): ceph-base-20.2.0-9.ge2ce8426bc5.el9.cl 6.4 MB/s | 5.9 MB 00:00 2026-04-01T13:53:24.561 INFO:teuthology.orchestra.run.vm09.stdout:(10/152): ceph-osd-20.2.0-9.ge2ce8426bc5.el9.cl 16 MB/s | 17 MB 00:01 2026-04-01T13:53:24.579 INFO:teuthology.orchestra.run.vm06.stdout:(5/152): ceph-mds-20.2.0-9.ge2ce8426bc5.el9.cly 9.4 MB/s | 2.3 MB 00:00 2026-04-01T13:53:24.646 INFO:teuthology.orchestra.run.vm09.stdout:(11/152): libcephfs-daemon-20.2.0-9.ge2ce8426bc 420 kB/s | 35 kB 00:00 2026-04-01T13:53:24.646 INFO:teuthology.orchestra.run.vm08.stdout:(8/152): ceph-mon-20.2.0-9.ge2ce8426bc5.el9.cly 5.0 MB/s | 5.0 MB 00:01 2026-04-01T13:53:24.727 INFO:teuthology.orchestra.run.vm06.stdout:(6/152): ceph-mgr-20.2.0-9.ge2ce8426bc5.el9.cly 4.9 MB/s | 961 kB 00:00 2026-04-01T13:53:24.735 INFO:teuthology.orchestra.run.vm09.stdout:(12/152): libcephfs-devel-20.2.0-9.ge2ce8426bc5 385 kB/s | 34 kB 00:00 2026-04-01T13:53:24.821 INFO:teuthology.orchestra.run.vm09.stdout:(13/152): libcephfs-proxy2-20.2.0-9.ge2ce8426bc 284 kB/s | 24 kB 00:00 2026-04-01T13:53:24.944 INFO:teuthology.orchestra.run.vm06.stdout:(7/152): ceph-common-20.2.0-9.ge2ce8426bc5.el9. 18 MB/s | 24 MB 00:01 2026-04-01T13:53:24.968 INFO:teuthology.orchestra.run.vm09.stdout:(14/152): libcephfs2-20.2.0-9.ge2ce8426bc5.el9. 5.8 MB/s | 868 kB 00:00 2026-04-01T13:53:24.983 INFO:teuthology.orchestra.run.vm08.stdout:(9/152): ceph-selinux-20.2.0-9.ge2ce8426bc5.el9 75 kB/s | 25 kB 00:00 2026-04-01T13:53:25.034 INFO:teuthology.orchestra.run.vm09.stdout:(15/152): libcephsqlite-20.2.0-9.ge2ce8426bc5.e 2.4 MB/s | 164 kB 00:00 2026-04-01T13:53:25.084 INFO:teuthology.orchestra.run.vm09.stdout:(16/152): librados-devel-20.2.0-9.ge2ce8426bc5. 2.5 MB/s | 126 kB 00:00 2026-04-01T13:53:25.236 INFO:teuthology.orchestra.run.vm09.stdout:(17/152): ceph-radosgw-20.2.0-9.ge2ce8426bc5.el 17 MB/s | 24 MB 00:01 2026-04-01T13:53:25.238 INFO:teuthology.orchestra.run.vm09.stdout:(18/152): libradosstriper1-20.2.0-9.ge2ce8426bc 1.6 MB/s | 250 kB 00:00 2026-04-01T13:53:25.304 INFO:teuthology.orchestra.run.vm09.stdout:(19/152): python3-ceph-argparse-20.2.0-9.ge2ce8 691 kB/s | 45 kB 00:00 2026-04-01T13:53:25.447 INFO:teuthology.orchestra.run.vm08.stdout:(10/152): ceph-osd-20.2.0-9.ge2ce8426bc5.el9.cl 9.8 MB/s | 17 MB 00:01 2026-04-01T13:53:25.467 INFO:teuthology.orchestra.run.vm09.stdout:(20/152): python3-ceph-common-20.2.0-9.ge2ce842 1.0 MB/s | 163 kB 00:00 2026-04-01T13:53:25.541 INFO:teuthology.orchestra.run.vm08.stdout:(11/152): ceph-radosgw-20.2.0-9.ge2ce8426bc5.el 16 MB/s | 24 MB 00:01 2026-04-01T13:53:25.542 INFO:teuthology.orchestra.run.vm08.stdout:(12/152): libcephfs-daemon-20.2.0-9.ge2ce8426bc 373 kB/s | 35 kB 00:00 2026-04-01T13:53:25.546 INFO:teuthology.orchestra.run.vm09.stdout:(21/152): python3-cephfs-20.2.0-9.ge2ce8426bc5. 2.0 MB/s | 163 kB 00:00 2026-04-01T13:53:25.608 INFO:teuthology.orchestra.run.vm08.stdout:(13/152): libcephfs-proxy2-20.2.0-9.ge2ce8426bc 374 kB/s | 24 kB 00:00 2026-04-01T13:53:25.706 INFO:teuthology.orchestra.run.vm06.stdout:(8/152): ceph-mon-20.2.0-9.ge2ce8426bc5.el9.cly 4.5 MB/s | 5.0 MB 00:01 2026-04-01T13:53:25.734 INFO:teuthology.orchestra.run.vm09.stdout:(22/152): python3-rados-20.2.0-9.ge2ce8426bc5.e 1.6 MB/s | 316 kB 00:00 2026-04-01T13:53:25.816 INFO:teuthology.orchestra.run.vm09.stdout:(23/152): python3-rbd-20.2.0-9.ge2ce8426bc5.el9 3.6 MB/s | 304 kB 00:00 2026-04-01T13:53:25.819 INFO:teuthology.orchestra.run.vm08.stdout:(14/152): libcephfs2-20.2.0-9.ge2ce8426bc5.el9. 4.0 MB/s | 868 kB 00:00 2026-04-01T13:53:25.821 INFO:teuthology.orchestra.run.vm08.stdout:(15/152): libcephfs-devel-20.2.0-9.ge2ce8426bc5 123 kB/s | 34 kB 00:00 2026-04-01T13:53:25.822 INFO:teuthology.orchestra.run.vm06.stdout:(9/152): ceph-selinux-20.2.0-9.ge2ce8426bc5.el9 217 kB/s | 25 kB 00:00 2026-04-01T13:53:25.897 INFO:teuthology.orchestra.run.vm08.stdout:(16/152): libcephsqlite-20.2.0-9.ge2ce8426bc5.e 2.1 MB/s | 164 kB 00:00 2026-04-01T13:53:25.902 INFO:teuthology.orchestra.run.vm09.stdout:(24/152): python3-rgw-20.2.0-9.ge2ce8426bc5.el9 1.1 MB/s | 99 kB 00:00 2026-04-01T13:53:25.946 INFO:teuthology.orchestra.run.vm09.stdout:(25/152): rbd-fuse-20.2.0-9.ge2ce8426bc5.el9.cl 2.0 MB/s | 91 kB 00:00 2026-04-01T13:53:25.988 INFO:teuthology.orchestra.run.vm08.stdout:(17/152): libradosstriper1-20.2.0-9.ge2ce8426bc 2.7 MB/s | 250 kB 00:00 2026-04-01T13:53:26.029 INFO:teuthology.orchestra.run.vm08.stdout:(18/152): librados-devel-20.2.0-9.ge2ce8426bc5. 604 kB/s | 126 kB 00:00 2026-04-01T13:53:26.134 INFO:teuthology.orchestra.run.vm08.stdout:(19/152): python3-ceph-argparse-20.2.0-9.ge2ce8 433 kB/s | 45 kB 00:00 2026-04-01T13:53:26.212 INFO:teuthology.orchestra.run.vm08.stdout:(20/152): python3-ceph-common-20.2.0-9.ge2ce842 2.0 MB/s | 163 kB 00:00 2026-04-01T13:53:26.303 INFO:teuthology.orchestra.run.vm06.stdout:(10/152): ceph-osd-20.2.0-9.ge2ce8426bc5.el9.cl 11 MB/s | 17 MB 00:01 2026-04-01T13:53:26.305 INFO:teuthology.orchestra.run.vm08.stdout:(21/152): python3-cephfs-20.2.0-9.ge2ce8426bc5. 1.7 MB/s | 163 kB 00:00 2026-04-01T13:53:26.343 INFO:teuthology.orchestra.run.vm06.stdout:(11/152): libcephfs-daemon-20.2.0-9.ge2ce8426bc 891 kB/s | 35 kB 00:00 2026-04-01T13:53:26.394 INFO:teuthology.orchestra.run.vm08.stdout:(22/152): python3-rados-20.2.0-9.ge2ce8426bc5.e 3.5 MB/s | 316 kB 00:00 2026-04-01T13:53:26.409 INFO:teuthology.orchestra.run.vm06.stdout:(12/152): libcephfs-devel-20.2.0-9.ge2ce8426bc5 519 kB/s | 34 kB 00:00 2026-04-01T13:53:26.455 INFO:teuthology.orchestra.run.vm08.stdout:(23/152): python3-rbd-20.2.0-9.ge2ce8426bc5.el9 4.9 MB/s | 304 kB 00:00 2026-04-01T13:53:26.499 INFO:teuthology.orchestra.run.vm08.stdout:(24/152): python3-rgw-20.2.0-9.ge2ce8426bc5.el9 2.2 MB/s | 99 kB 00:00 2026-04-01T13:53:26.540 INFO:teuthology.orchestra.run.vm08.stdout:(25/152): rbd-fuse-20.2.0-9.ge2ce8426bc5.el9.cl 2.2 MB/s | 91 kB 00:00 2026-04-01T13:53:26.850 INFO:teuthology.orchestra.run.vm08.stdout:(26/152): librgw2-20.2.0-9.ge2ce8426bc5.el9.cly 7.4 MB/s | 6.4 MB 00:00 2026-04-01T13:53:26.851 INFO:teuthology.orchestra.run.vm09.stdout:(26/152): librgw2-20.2.0-9.ge2ce8426bc5.el9.cly 3.9 MB/s | 6.4 MB 00:01 2026-04-01T13:53:26.908 INFO:teuthology.orchestra.run.vm09.stdout:(27/152): rbd-nbd-20.2.0-9.ge2ce8426bc5.el9.cly 3.1 MB/s | 180 kB 00:00 2026-04-01T13:53:26.914 INFO:teuthology.orchestra.run.vm08.stdout:(27/152): rbd-nbd-20.2.0-9.ge2ce8426bc5.el9.cly 2.8 MB/s | 180 kB 00:00 2026-04-01T13:53:26.972 INFO:teuthology.orchestra.run.vm09.stdout:(28/152): ceph-grafana-dashboards-20.2.0-9.ge2c 670 kB/s | 43 kB 00:00 2026-04-01T13:53:26.973 INFO:teuthology.orchestra.run.vm08.stdout:(28/152): ceph-grafana-dashboards-20.2.0-9.ge2c 734 kB/s | 43 kB 00:00 2026-04-01T13:53:27.023 INFO:teuthology.orchestra.run.vm09.stdout:(29/152): ceph-mgr-cephadm-20.2.0-9.ge2ce8426bc 3.4 MB/s | 173 kB 00:00 2026-04-01T13:53:27.031 INFO:teuthology.orchestra.run.vm08.stdout:(29/152): ceph-mgr-cephadm-20.2.0-9.ge2ce8426bc 2.9 MB/s | 173 kB 00:00 2026-04-01T13:53:27.111 INFO:teuthology.orchestra.run.vm06.stdout:(13/152): libcephfs-proxy2-20.2.0-9.ge2ce8426bc 34 kB/s | 24 kB 00:00 2026-04-01T13:53:27.290 INFO:teuthology.orchestra.run.vm09.stdout:(30/152): rbd-mirror-20.2.0-9.ge2ce8426bc5.el9. 2.2 MB/s | 2.9 MB 00:01 2026-04-01T13:53:27.426 INFO:teuthology.orchestra.run.vm06.stdout:(14/152): libcephfs2-20.2.0-9.ge2ce8426bc5.el9. 2.7 MB/s | 868 kB 00:00 2026-04-01T13:53:27.485 INFO:teuthology.orchestra.run.vm06.stdout:(15/152): libcephsqlite-20.2.0-9.ge2ce8426bc5.e 2.7 MB/s | 164 kB 00:00 2026-04-01T13:53:27.551 INFO:teuthology.orchestra.run.vm08.stdout:(30/152): rbd-mirror-20.2.0-9.ge2ce8426bc5.el9. 2.9 MB/s | 2.9 MB 00:01 2026-04-01T13:53:27.815 INFO:teuthology.orchestra.run.vm06.stdout:(16/152): ceph-radosgw-20.2.0-9.ge2ce8426bc5.el 8.3 MB/s | 24 MB 00:02 2026-04-01T13:53:27.817 INFO:teuthology.orchestra.run.vm06.stdout:(17/152): librados-devel-20.2.0-9.ge2ce8426bc5. 378 kB/s | 126 kB 00:00 2026-04-01T13:53:27.940 INFO:teuthology.orchestra.run.vm06.stdout:(18/152): libradosstriper1-20.2.0-9.ge2ce8426bc 2.0 MB/s | 250 kB 00:00 2026-04-01T13:53:28.168 INFO:teuthology.orchestra.run.vm08.stdout:(31/152): ceph-mgr-diskprediction-local-20.2.0- 12 MB/s | 7.4 MB 00:00 2026-04-01T13:53:28.241 INFO:teuthology.orchestra.run.vm08.stdout:(32/152): ceph-mgr-k8sevents-20.2.0-9.ge2ce8426 305 kB/s | 22 kB 00:00 2026-04-01T13:53:28.315 INFO:teuthology.orchestra.run.vm08.stdout:(33/152): ceph-mgr-modules-core-20.2.0-9.ge2ce8 3.8 MB/s | 289 kB 00:00 2026-04-01T13:53:28.383 INFO:teuthology.orchestra.run.vm08.stdout:(34/152): ceph-mgr-rook-20.2.0-9.ge2ce8426bc5.e 741 kB/s | 50 kB 00:00 2026-04-01T13:53:28.418 INFO:teuthology.orchestra.run.vm08.stdout:(35/152): ceph-prometheus-alerts-20.2.0-9.ge2ce 500 kB/s | 17 kB 00:00 2026-04-01T13:53:28.435 INFO:teuthology.orchestra.run.vm06.stdout:(19/152): librgw2-20.2.0-9.ge2ce8426bc5.el9.cly 10 MB/s | 6.4 MB 00:00 2026-04-01T13:53:28.487 INFO:teuthology.orchestra.run.vm08.stdout:(36/152): ceph-volume-20.2.0-9.ge2ce8426bc5.el9 4.3 MB/s | 297 kB 00:00 2026-04-01T13:53:28.500 INFO:teuthology.orchestra.run.vm06.stdout:(20/152): python3-ceph-common-20.2.0-9.ge2ce842 2.5 MB/s | 163 kB 00:00 2026-04-01T13:53:28.562 INFO:teuthology.orchestra.run.vm06.stdout:(21/152): python3-ceph-argparse-20.2.0-9.ge2ce8 72 kB/s | 45 kB 00:00 2026-04-01T13:53:28.574 INFO:teuthology.orchestra.run.vm06.stdout:(22/152): python3-cephfs-20.2.0-9.ge2ce8426bc5. 2.2 MB/s | 163 kB 00:00 2026-04-01T13:53:28.610 INFO:teuthology.orchestra.run.vm08.stdout:(37/152): cephadm-20.2.0-9.ge2ce8426bc5.el9.cly 8.0 MB/s | 1.0 MB 00:00 2026-04-01T13:53:28.631 INFO:teuthology.orchestra.run.vm09.stdout:(31/152): ceph-mgr-diskprediction-local-20.2.0- 5.5 MB/s | 7.4 MB 00:01 2026-04-01T13:53:28.636 INFO:teuthology.orchestra.run.vm06.stdout:(23/152): python3-rbd-20.2.0-9.ge2ce8426bc5.el9 4.8 MB/s | 304 kB 00:00 2026-04-01T13:53:28.644 INFO:teuthology.orchestra.run.vm06.stdout:(24/152): python3-rados-20.2.0-9.ge2ce8426bc5.e 3.8 MB/s | 316 kB 00:00 2026-04-01T13:53:28.662 INFO:teuthology.orchestra.run.vm08.stdout:(38/152): abseil-cpp-20211102.0-4.el9.x86_64.rp 10 MB/s | 551 kB 00:00 2026-04-01T13:53:28.669 INFO:teuthology.orchestra.run.vm09.stdout:(32/152): ceph-mgr-k8sevents-20.2.0-9.ge2ce8426 597 kB/s | 22 kB 00:00 2026-04-01T13:53:28.678 INFO:teuthology.orchestra.run.vm06.stdout:(25/152): python3-rgw-20.2.0-9.ge2ce8426bc5.el9 2.3 MB/s | 99 kB 00:00 2026-04-01T13:53:28.684 INFO:teuthology.orchestra.run.vm06.stdout:(26/152): rbd-fuse-20.2.0-9.ge2ce8426bc5.el9.cl 2.3 MB/s | 91 kB 00:00 2026-04-01T13:53:28.686 INFO:teuthology.orchestra.run.vm08.stdout:(39/152): gperftools-libs-2.9.1-3.el9.x86_64.rp 12 MB/s | 308 kB 00:00 2026-04-01T13:53:28.693 INFO:teuthology.orchestra.run.vm08.stdout:(40/152): grpc-data-1.46.7-10.el9.noarch.rpm 3.2 MB/s | 19 kB 00:00 2026-04-01T13:53:28.780 INFO:teuthology.orchestra.run.vm09.stdout:(33/152): ceph-mgr-modules-core-20.2.0-9.ge2ce8 2.5 MB/s | 289 kB 00:00 2026-04-01T13:53:28.799 INFO:teuthology.orchestra.run.vm06.stdout:(27/152): rbd-nbd-20.2.0-9.ge2ce8426bc5.el9.cly 1.5 MB/s | 180 kB 00:00 2026-04-01T13:53:28.827 INFO:teuthology.orchestra.run.vm09.stdout:(34/152): ceph-mgr-rook-20.2.0-9.ge2ce8426bc5.e 1.1 MB/s | 50 kB 00:00 2026-04-01T13:53:28.873 INFO:teuthology.orchestra.run.vm06.stdout:(28/152): ceph-grafana-dashboards-20.2.0-9.ge2c 585 kB/s | 43 kB 00:00 2026-04-01T13:53:28.875 INFO:teuthology.orchestra.run.vm09.stdout:(35/152): ceph-prometheus-alerts-20.2.0-9.ge2ce 357 kB/s | 17 kB 00:00 2026-04-01T13:53:28.978 INFO:teuthology.orchestra.run.vm06.stdout:(29/152): ceph-mgr-cephadm-20.2.0-9.ge2ce8426bc 1.6 MB/s | 173 kB 00:00 2026-04-01T13:53:29.058 INFO:teuthology.orchestra.run.vm08.stdout:(41/152): libarrow-9.0.0-15.el9.x86_64.rpm 12 MB/s | 4.4 MB 00:00 2026-04-01T13:53:29.058 INFO:teuthology.orchestra.run.vm06.stdout:(30/152): rbd-mirror-20.2.0-9.ge2ce8426bc5.el9. 7.7 MB/s | 2.9 MB 00:00 2026-04-01T13:53:29.078 INFO:teuthology.orchestra.run.vm08.stdout:(42/152): libarrow-doc-9.0.0-15.el9.noarch.rpm 1.2 MB/s | 25 kB 00:00 2026-04-01T13:53:29.081 INFO:teuthology.orchestra.run.vm08.stdout:(43/152): liboath-2.6.12-1.el9.x86_64.rpm 16 MB/s | 49 kB 00:00 2026-04-01T13:53:29.142 INFO:teuthology.orchestra.run.vm08.stdout:(44/152): libunwind-1.6.2-1.el9.x86_64.rpm 1.1 MB/s | 67 kB 00:00 2026-04-01T13:53:29.190 INFO:teuthology.orchestra.run.vm08.stdout:(45/152): luarocks-3.9.2-5.el9.noarch.rpm 3.2 MB/s | 151 kB 00:00 2026-04-01T13:53:29.190 INFO:teuthology.orchestra.run.vm09.stdout:(36/152): ceph-test-20.2.0-9.ge2ce8426bc5.el9.c 17 MB/s | 85 MB 00:05 2026-04-01T13:53:29.192 INFO:teuthology.orchestra.run.vm09.stdout:(37/152): ceph-volume-20.2.0-9.ge2ce8426bc5.el9 937 kB/s | 297 kB 00:00 2026-04-01T13:53:29.284 INFO:teuthology.orchestra.run.vm08.stdout:(46/152): parquet-libs-9.0.0-15.el9.x86_64.rpm 8.8 MB/s | 838 kB 00:00 2026-04-01T13:53:29.284 INFO:teuthology.orchestra.run.vm09.stdout:(38/152): ceph-mgr-dashboard-20.2.0-9.ge2ce8426 6.7 MB/s | 15 MB 00:02 2026-04-01T13:53:29.318 INFO:teuthology.orchestra.run.vm09.stdout:(39/152): gperftools-libs-2.9.1-3.el9.x86_64.rp 9.0 MB/s | 308 kB 00:00 2026-04-01T13:53:29.329 INFO:teuthology.orchestra.run.vm09.stdout:(40/152): grpc-data-1.46.7-10.el9.noarch.rpm 1.8 MB/s | 19 kB 00:00 2026-04-01T13:53:29.329 INFO:teuthology.orchestra.run.vm08.stdout:(47/152): ceph-mgr-dashboard-20.2.0-9.ge2ce8426 6.6 MB/s | 15 MB 00:02 2026-04-01T13:53:29.391 INFO:teuthology.orchestra.run.vm08.stdout:(48/152): python3-autocommand-2.2.2-8.el9.noarc 480 kB/s | 29 kB 00:00 2026-04-01T13:53:29.397 INFO:teuthology.orchestra.run.vm09.stdout:(41/152): abseil-cpp-20211102.0-4.el9.x86_64.rp 2.6 MB/s | 551 kB 00:00 2026-04-01T13:53:29.400 INFO:teuthology.orchestra.run.vm08.stdout:(49/152): python3-asyncssh-2.13.2-5.el9.noarch. 4.6 MB/s | 548 kB 00:00 2026-04-01T13:53:29.401 INFO:teuthology.orchestra.run.vm09.stdout:(42/152): libarrow-doc-9.0.0-15.el9.noarch.rpm 7.2 MB/s | 25 kB 00:00 2026-04-01T13:53:29.402 INFO:teuthology.orchestra.run.vm08.stdout:(50/152): python3-backports-tarfile-1.2.0-1.el9 5.1 MB/s | 60 kB 00:00 2026-04-01T13:53:29.405 INFO:teuthology.orchestra.run.vm08.stdout:(51/152): python3-bcrypt-3.2.2-1.el9.x86_64.rpm 8.6 MB/s | 43 kB 00:00 2026-04-01T13:53:29.407 INFO:teuthology.orchestra.run.vm09.stdout:(43/152): cephadm-20.2.0-9.ge2ce8426bc5.el9.cly 4.5 MB/s | 1.0 MB 00:00 2026-04-01T13:53:29.408 INFO:teuthology.orchestra.run.vm08.stdout:(52/152): python3-certifi-2023.05.07-4.el9.noar 5.9 MB/s | 14 kB 00:00 2026-04-01T13:53:29.408 INFO:teuthology.orchestra.run.vm09.stdout:(44/152): liboath-2.6.12-1.el9.x86_64.rpm 6.8 MB/s | 49 kB 00:00 2026-04-01T13:53:29.409 INFO:teuthology.orchestra.run.vm08.stdout:(53/152): python3-cachetools-4.2.4-1.el9.noarch 5.1 MB/s | 32 kB 00:00 2026-04-01T13:53:29.420 INFO:teuthology.orchestra.run.vm09.stdout:(45/152): libunwind-1.6.2-1.el9.x86_64.rpm 5.1 MB/s | 67 kB 00:00 2026-04-01T13:53:29.428 INFO:teuthology.orchestra.run.vm08.stdout:(54/152): python3-cheroot-10.0.1-5.el9.noarch.r 8.7 MB/s | 173 kB 00:00 2026-04-01T13:53:29.429 INFO:teuthology.orchestra.run.vm09.stdout:(46/152): luarocks-3.9.2-5.el9.noarch.rpm 7.2 MB/s | 151 kB 00:00 2026-04-01T13:53:29.457 INFO:teuthology.orchestra.run.vm09.stdout:(47/152): parquet-libs-9.0.0-15.el9.x86_64.rpm 23 MB/s | 838 kB 00:00 2026-04-01T13:53:29.461 INFO:teuthology.orchestra.run.vm09.stdout:(48/152): python3-autocommand-2.2.2-8.el9.noarc 9.0 MB/s | 29 kB 00:00 2026-04-01T13:53:29.464 INFO:teuthology.orchestra.run.vm08.stdout:(55/152): python3-google-auth-2.45.0-1.el9.noar 7.0 MB/s | 254 kB 00:00 2026-04-01T13:53:29.464 INFO:teuthology.orchestra.run.vm09.stdout:(49/152): python3-backports-tarfile-1.2.0-1.el9 17 MB/s | 60 kB 00:00 2026-04-01T13:53:29.469 INFO:teuthology.orchestra.run.vm09.stdout:(50/152): python3-bcrypt-3.2.2-1.el9.x86_64.rpm 11 MB/s | 43 kB 00:00 2026-04-01T13:53:29.474 INFO:teuthology.orchestra.run.vm09.stdout:(51/152): python3-cachetools-4.2.4-1.el9.noarch 6.1 MB/s | 32 kB 00:00 2026-04-01T13:53:29.476 INFO:teuthology.orchestra.run.vm08.stdout:(56/152): python3-cherrypy-18.10.0-5.el9.noarch 4.3 MB/s | 290 kB 00:00 2026-04-01T13:53:29.477 INFO:teuthology.orchestra.run.vm09.stdout:(52/152): python3-certifi-2023.05.07-4.el9.noar 5.4 MB/s | 14 kB 00:00 2026-04-01T13:53:29.483 INFO:teuthology.orchestra.run.vm09.stdout:(53/152): python3-cheroot-10.0.1-5.el9.noarch.r 33 MB/s | 173 kB 00:00 2026-04-01T13:53:29.491 INFO:teuthology.orchestra.run.vm09.stdout:(54/152): python3-cherrypy-18.10.0-5.el9.noarch 35 MB/s | 290 kB 00:00 2026-04-01T13:53:29.498 INFO:teuthology.orchestra.run.vm08.stdout:(57/152): python3-grpcio-tools-1.46.7-10.el9.x8 6.3 MB/s | 144 kB 00:00 2026-04-01T13:53:29.502 INFO:teuthology.orchestra.run.vm09.stdout:(55/152): python3-google-auth-2.45.0-1.el9.noar 25 MB/s | 254 kB 00:00 2026-04-01T13:53:29.517 INFO:teuthology.orchestra.run.vm09.stdout:(56/152): python3-asyncssh-2.13.2-5.el9.noarch. 6.1 MB/s | 548 kB 00:00 2026-04-01T13:53:29.525 INFO:teuthology.orchestra.run.vm08.stdout:(58/152): python3-influxdb-5.3.1-1.el9.noarch.r 5.2 MB/s | 139 kB 00:00 2026-04-01T13:53:29.538 INFO:teuthology.orchestra.run.vm08.stdout:(59/152): python3-isodate-0.6.1-3.el9.noarch.rp 4.4 MB/s | 56 kB 00:00 2026-04-01T13:53:29.543 INFO:teuthology.orchestra.run.vm09.stdout:(57/152): python3-grpcio-tools-1.46.7-10.el9.x8 5.5 MB/s | 144 kB 00:00 2026-04-01T13:53:29.543 INFO:teuthology.orchestra.run.vm08.stdout:(60/152): python3-jaraco-8.2.1-3.el9.noarch.rpm 2.0 MB/s | 11 kB 00:00 2026-04-01T13:53:29.549 INFO:teuthology.orchestra.run.vm08.stdout:(61/152): python3-jaraco-classes-3.2.1-5.el9.no 2.8 MB/s | 18 kB 00:00 2026-04-01T13:53:29.556 INFO:teuthology.orchestra.run.vm08.stdout:(62/152): python3-jaraco-collections-3.0.0-8.el 3.8 MB/s | 23 kB 00:00 2026-04-01T13:53:29.556 INFO:teuthology.orchestra.run.vm09.stdout:(58/152): python3-grpcio-1.46.7-10.el9.x86_64.r 38 MB/s | 2.0 MB 00:00 2026-04-01T13:53:29.560 INFO:teuthology.orchestra.run.vm09.stdout:(59/152): python3-isodate-0.6.1-3.el9.noarch.rp 13 MB/s | 56 kB 00:00 2026-04-01T13:53:29.560 INFO:teuthology.orchestra.run.vm08.stdout:(63/152): python3-jaraco-context-6.0.1-3.el9.no 4.5 MB/s | 20 kB 00:00 2026-04-01T13:53:29.564 INFO:teuthology.orchestra.run.vm08.stdout:(64/152): python3-jaraco-functools-3.5.0-2.el9. 6.3 MB/s | 19 kB 00:00 2026-04-01T13:53:29.567 INFO:teuthology.orchestra.run.vm08.stdout:(65/152): python3-jaraco-text-4.0.0-2.el9.noarc 8.4 MB/s | 26 kB 00:00 2026-04-01T13:53:29.573 INFO:teuthology.orchestra.run.vm09.stdout:(60/152): libarrow-9.0.0-15.el9.x86_64.rpm 18 MB/s | 4.4 MB 00:00 2026-04-01T13:53:29.574 INFO:teuthology.orchestra.run.vm09.stdout:(61/152): python3-influxdb-5.3.1-1.el9.noarch.r 4.3 MB/s | 139 kB 00:00 2026-04-01T13:53:29.575 INFO:teuthology.orchestra.run.vm09.stdout:(62/152): python3-jaraco-8.2.1-3.el9.noarch.rpm 720 kB/s | 11 kB 00:00 2026-04-01T13:53:29.578 INFO:teuthology.orchestra.run.vm09.stdout:(63/152): python3-jaraco-classes-3.2.1-5.el9.no 5.5 MB/s | 18 kB 00:00 2026-04-01T13:53:29.578 INFO:teuthology.orchestra.run.vm09.stdout:(64/152): python3-jaraco-collections-3.0.0-8.el 6.1 MB/s | 23 kB 00:00 2026-04-01T13:53:29.579 INFO:teuthology.orchestra.run.vm09.stdout:(65/152): python3-jaraco-context-6.0.1-3.el9.no 4.8 MB/s | 20 kB 00:00 2026-04-01T13:53:29.581 INFO:teuthology.orchestra.run.vm09.stdout:(66/152): python3-jaraco-functools-3.5.0-2.el9. 6.8 MB/s | 19 kB 00:00 2026-04-01T13:53:29.583 INFO:teuthology.orchestra.run.vm09.stdout:(67/152): python3-jaraco-text-4.0.0-2.el9.noarc 6.4 MB/s | 26 kB 00:00 2026-04-01T13:53:29.588 INFO:teuthology.orchestra.run.vm09.stdout:(68/152): python3-msgpack-1.0.3-2.el9.x86_64.rp 17 MB/s | 86 kB 00:00 2026-04-01T13:53:29.593 INFO:teuthology.orchestra.run.vm09.stdout:(69/152): python3-more-itertools-8.12.0-2.el9.n 6.6 MB/s | 79 kB 00:00 2026-04-01T13:53:29.595 INFO:teuthology.orchestra.run.vm09.stdout:(70/152): python3-natsort-7.1.1-5.el9.noarch.rp 8.2 MB/s | 58 kB 00:00 2026-04-01T13:53:29.598 INFO:teuthology.orchestra.run.vm09.stdout:(71/152): python3-portend-3.1.0-2.el9.noarch.rp 3.3 MB/s | 16 kB 00:00 2026-04-01T13:53:29.601 INFO:teuthology.orchestra.run.vm09.stdout:(72/152): python3-pyOpenSSL-21.0.0-1.el9.noarch 15 MB/s | 90 kB 00:00 2026-04-01T13:53:29.609 INFO:teuthology.orchestra.run.vm09.stdout:(73/152): python3-kubernetes-26.1.0-3.el9.noarc 35 MB/s | 1.0 MB 00:00 2026-04-01T13:53:29.609 INFO:teuthology.orchestra.run.vm08.stdout:(66/152): python3-grpcio-1.46.7-10.el9.x86_64.r 14 MB/s | 2.0 MB 00:00 2026-04-01T13:53:29.610 INFO:teuthology.orchestra.run.vm09.stdout:(74/152): python3-repoze-lru-0.7-16.el9.noarch. 2.7 MB/s | 31 kB 00:00 2026-04-01T13:53:29.612 INFO:teuthology.orchestra.run.vm09.stdout:(75/152): python3-routes-2.5.1-5.el9.noarch.rpm 17 MB/s | 188 kB 00:00 2026-04-01T13:53:29.614 INFO:teuthology.orchestra.run.vm09.stdout:(76/152): python3-rsa-4.9-2.el9.noarch.rpm 13 MB/s | 59 kB 00:00 2026-04-01T13:53:29.615 INFO:teuthology.orchestra.run.vm08.stdout:(67/152): python3-more-itertools-8.12.0-2.el9.n 14 MB/s | 79 kB 00:00 2026-04-01T13:53:29.615 INFO:teuthology.orchestra.run.vm09.stdout:(77/152): python3-tempora-5.0.0-2.el9.noarch.rp 11 MB/s | 36 kB 00:00 2026-04-01T13:53:29.626 INFO:teuthology.orchestra.run.vm06.stdout:(31/152): ceph-mgr-diskprediction-local-20.2.0- 13 MB/s | 7.4 MB 00:00 2026-04-01T13:53:29.626 INFO:teuthology.orchestra.run.vm09.stdout:(78/152): python3-typing-extensions-4.15.0-1.el 6.8 MB/s | 86 kB 00:00 2026-04-01T13:53:29.626 INFO:teuthology.orchestra.run.vm08.stdout:(68/152): python3-msgpack-1.0.3-2.el9.x86_64.rp 7.8 MB/s | 86 kB 00:00 2026-04-01T13:53:29.628 INFO:teuthology.orchestra.run.vm09.stdout:(79/152): python3-saml-1.16.0-1.el9.noarch.rpm 6.8 MB/s | 125 kB 00:00 2026-04-01T13:53:29.629 INFO:teuthology.orchestra.run.vm09.stdout:(80/152): python3-websocket-client-1.2.3-2.el9. 6.3 MB/s | 90 kB 00:00 2026-04-01T13:53:29.631 INFO:teuthology.orchestra.run.vm08.stdout:(69/152): python3-natsort-7.1.1-5.el9.noarch.rp 14 MB/s | 58 kB 00:00 2026-04-01T13:53:29.631 INFO:teuthology.orchestra.run.vm09.stdout:(81/152): python3-xmlsec-1.3.13-1.el9.x86_64.rp 11 MB/s | 48 kB 00:00 2026-04-01T13:53:29.632 INFO:teuthology.orchestra.run.vm09.stdout:(82/152): python3-xmltodict-0.12.0-15.el9.noarc 6.1 MB/s | 22 kB 00:00 2026-04-01T13:53:29.632 INFO:teuthology.orchestra.run.vm09.stdout:(83/152): python3-zc-lockfile-2.0-10.el9.noarch 7.4 MB/s | 20 kB 00:00 2026-04-01T13:53:29.635 INFO:teuthology.orchestra.run.vm08.stdout:(70/152): python3-portend-3.1.0-2.el9.noarch.rp 3.6 MB/s | 16 kB 00:00 2026-04-01T13:53:29.636 INFO:teuthology.orchestra.run.vm09.stdout:(84/152): re2-20211101-20.el9.x86_64.rpm 41 MB/s | 191 kB 00:00 2026-04-01T13:53:29.641 INFO:teuthology.orchestra.run.vm08.stdout:(71/152): python3-kubernetes-26.1.0-3.el9.noarc 14 MB/s | 1.0 MB 00:00 2026-04-01T13:53:29.644 INFO:teuthology.orchestra.run.vm08.stdout:(72/152): python3-pyOpenSSL-21.0.0-1.el9.noarch 11 MB/s | 90 kB 00:00 2026-04-01T13:53:29.645 INFO:teuthology.orchestra.run.vm08.stdout:(73/152): python3-repoze-lru-0.7-16.el9.noarch. 7.3 MB/s | 31 kB 00:00 2026-04-01T13:53:29.645 INFO:teuthology.orchestra.run.vm09.stdout:(85/152): s3cmd-2.4.0-1.el9.noarch.rpm 15 MB/s | 206 kB 00:00 2026-04-01T13:53:29.650 INFO:teuthology.orchestra.run.vm08.stdout:(74/152): python3-rsa-4.9-2.el9.noarch.rpm 12 MB/s | 59 kB 00:00 2026-04-01T13:53:29.652 INFO:teuthology.orchestra.run.vm08.stdout:(75/152): python3-routes-2.5.1-5.el9.noarch.rpm 23 MB/s | 188 kB 00:00 2026-04-01T13:53:29.657 INFO:teuthology.orchestra.run.vm08.stdout:(76/152): python3-tempora-5.0.0-2.el9.noarch.rp 7.1 MB/s | 36 kB 00:00 2026-04-01T13:53:29.661 INFO:teuthology.orchestra.run.vm08.stdout:(77/152): python3-saml-1.16.0-1.el9.noarch.rpm 11 MB/s | 125 kB 00:00 2026-04-01T13:53:29.664 INFO:teuthology.orchestra.run.vm08.stdout:(78/152): python3-typing-extensions-4.15.0-1.el 13 MB/s | 86 kB 00:00 2026-04-01T13:53:29.667 INFO:teuthology.orchestra.run.vm08.stdout:(79/152): python3-websocket-client-1.2.3-2.el9. 17 MB/s | 90 kB 00:00 2026-04-01T13:53:29.668 INFO:teuthology.orchestra.run.vm08.stdout:(80/152): python3-xmlsec-1.3.13-1.el9.x86_64.rp 12 MB/s | 48 kB 00:00 2026-04-01T13:53:29.669 INFO:teuthology.orchestra.run.vm08.stdout:(81/152): python3-xmltodict-0.12.0-15.el9.noarc 8.7 MB/s | 22 kB 00:00 2026-04-01T13:53:29.671 INFO:teuthology.orchestra.run.vm08.stdout:(82/152): python3-zc-lockfile-2.0-10.el9.noarch 6.7 MB/s | 20 kB 00:00 2026-04-01T13:53:29.675 INFO:teuthology.orchestra.run.vm09.stdout:(86/152): thrift-0.15.0-4.el9.x86_64.rpm 37 MB/s | 1.6 MB 00:00 2026-04-01T13:53:29.677 INFO:teuthology.orchestra.run.vm08.stdout:(83/152): re2-20211101-20.el9.x86_64.rpm 27 MB/s | 191 kB 00:00 2026-04-01T13:53:29.679 INFO:teuthology.orchestra.run.vm08.stdout:(84/152): s3cmd-2.4.0-1.el9.noarch.rpm 28 MB/s | 206 kB 00:00 2026-04-01T13:53:29.719 INFO:teuthology.orchestra.run.vm08.stdout:(85/152): thrift-0.15.0-4.el9.x86_64.rpm 37 MB/s | 1.6 MB 00:00 2026-04-01T13:53:29.742 INFO:teuthology.orchestra.run.vm06.stdout:(32/152): ceph-mgr-k8sevents-20.2.0-9.ge2ce8426 191 kB/s | 22 kB 00:00 2026-04-01T13:53:29.770 INFO:teuthology.orchestra.run.vm09.stdout:(87/152): bzip2-1.0.8-10.el9_5.x86_64.rpm 384 kB/s | 51 kB 00:00 2026-04-01T13:53:29.799 INFO:teuthology.orchestra.run.vm06.stdout:(33/152): ceph-mgr-modules-core-20.2.0-9.ge2ce8 5.1 MB/s | 289 kB 00:00 2026-04-01T13:53:29.813 INFO:teuthology.orchestra.run.vm08.stdout:(86/152): bzip2-1.0.8-10.el9_5.x86_64.rpm 385 kB/s | 51 kB 00:00 2026-04-01T13:53:29.821 INFO:teuthology.orchestra.run.vm09.stdout:(88/152): c-ares-1.19.1-2.el9_4.x86_64.rpm 627 kB/s | 110 kB 00:00 2026-04-01T13:53:29.822 INFO:teuthology.orchestra.run.vm09.stdout:(89/152): fuse-2.9.9-17.el9.x86_64.rpm 1.5 MB/s | 78 kB 00:00 2026-04-01T13:53:29.832 INFO:teuthology.orchestra.run.vm06.stdout:(34/152): ceph-mgr-rook-20.2.0-9.ge2ce8426bc5.e 1.5 MB/s | 50 kB 00:00 2026-04-01T13:53:29.848 INFO:teuthology.orchestra.run.vm09.stdout:(90/152): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-01T13:53:29.849 INFO:teuthology.orchestra.run.vm09.stdout:(91/152): libconfig-1.7.2-9.el9.x86_64.rpm 2.5 MB/s | 71 kB 00:00 2026-04-01T13:53:29.860 INFO:teuthology.orchestra.run.vm06.stdout:(35/152): ceph-prometheus-alerts-20.2.0-9.ge2ce 611 kB/s | 17 kB 00:00 2026-04-01T13:53:29.864 INFO:teuthology.orchestra.run.vm09.stdout:(92/152): cryptsetup-2.7.2-4.el9.x86_64.rpm 1.6 MB/s | 310 kB 00:00 2026-04-01T13:53:29.879 INFO:teuthology.orchestra.run.vm08.stdout:(87/152): c-ares-1.19.1-2.el9_4.x86_64.rpm 689 kB/s | 110 kB 00:00 2026-04-01T13:53:29.892 INFO:teuthology.orchestra.run.vm08.stdout:(88/152): cryptsetup-2.7.2-4.el9.x86_64.rpm 3.8 MB/s | 310 kB 00:00 2026-04-01T13:53:29.892 INFO:teuthology.orchestra.run.vm09.stdout:(93/152): lmdb-libs-0.9.29-3.el9.x86_64.rpm 2.2 MB/s | 60 kB 00:00 2026-04-01T13:53:29.905 INFO:teuthology.orchestra.run.vm06.stdout:(36/152): ceph-volume-20.2.0-9.ge2ce8426bc5.el9 6.5 MB/s | 297 kB 00:00 2026-04-01T13:53:29.906 INFO:teuthology.orchestra.run.vm09.stdout:(94/152): libquadmath-11.5.0-11.el9.x86_64.rpm 3.2 MB/s | 184 kB 00:00 2026-04-01T13:53:29.909 INFO:teuthology.orchestra.run.vm08.stdout:(89/152): fuse-2.9.9-17.el9.x86_64.rpm 2.6 MB/s | 78 kB 00:00 2026-04-01T13:53:29.919 INFO:teuthology.orchestra.run.vm09.stdout:(95/152): mailcap-2.1.49-5.el9.0.2.noarch.rpm 1.2 MB/s | 32 kB 00:00 2026-04-01T13:53:29.919 INFO:teuthology.orchestra.run.vm08.stdout:(90/152): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-01T13:53:29.930 INFO:teuthology.orchestra.run.vm09.stdout:(96/152): libgfortran-11.5.0-11.el9.x86_64.rpm 9.5 MB/s | 794 kB 00:00 2026-04-01T13:53:29.937 INFO:teuthology.orchestra.run.vm08.stdout:(91/152): libconfig-1.7.2-9.el9.x86_64.rpm 2.5 MB/s | 71 kB 00:00 2026-04-01T13:53:29.948 INFO:teuthology.orchestra.run.vm09.stdout:(97/152): pciutils-3.7.0-7.el9.x86_64.rpm 3.1 MB/s | 92 kB 00:00 2026-04-01T13:53:29.962 INFO:teuthology.orchestra.run.vm09.stdout:(98/152): python3-cffi-1.14.5-5.el9.x86_64.rpm 7.4 MB/s | 241 kB 00:00 2026-04-01T13:53:29.983 INFO:teuthology.orchestra.run.vm08.stdout:(92/152): libgfortran-11.5.0-11.el9.x86_64.rpm 12 MB/s | 794 kB 00:00 2026-04-01T13:53:29.990 INFO:teuthology.orchestra.run.vm09.stdout:(99/152): python3-ply-3.11-14.el9.0.1.noarch.rp 3.6 MB/s | 103 kB 00:00 2026-04-01T13:53:29.999 INFO:teuthology.orchestra.run.vm08.stdout:(93/152): libquadmath-11.5.0-11.el9.x86_64.rpm 2.9 MB/s | 184 kB 00:00 2026-04-01T13:53:30.011 INFO:teuthology.orchestra.run.vm08.stdout:(94/152): lmdb-libs-0.9.29-3.el9.x86_64.rpm 2.1 MB/s | 60 kB 00:00 2026-04-01T13:53:30.015 INFO:teuthology.orchestra.run.vm06.stdout:(37/152): cephadm-20.2.0-9.ge2ce8426bc5.el9.cly 8.9 MB/s | 1.0 MB 00:00 2026-04-01T13:53:30.017 INFO:teuthology.orchestra.run.vm09.stdout:(100/152): python3-pycparser-2.20-6.el9.noarch. 4.4 MB/s | 124 kB 00:00 2026-04-01T13:53:30.026 INFO:teuthology.orchestra.run.vm08.stdout:(95/152): mailcap-2.1.49-5.el9.0.2.noarch.rpm 1.2 MB/s | 32 kB 00:00 2026-04-01T13:53:30.033 INFO:teuthology.orchestra.run.vm06.stdout:(38/152): abseil-cpp-20211102.0-4.el9.x86_64.rp 31 MB/s | 551 kB 00:00 2026-04-01T13:53:30.034 INFO:teuthology.orchestra.run.vm09.stdout:(101/152): python3-cryptography-36.0.1-5.el9_6. 14 MB/s | 1.2 MB 00:00 2026-04-01T13:53:30.043 INFO:teuthology.orchestra.run.vm06.stdout:(39/152): gperftools-libs-2.9.1-3.el9.x86_64.rp 34 MB/s | 308 kB 00:00 2026-04-01T13:53:30.048 INFO:teuthology.orchestra.run.vm06.stdout:(40/152): grpc-data-1.46.7-10.el9.noarch.rpm 3.8 MB/s | 19 kB 00:00 2026-04-01T13:53:30.050 INFO:teuthology.orchestra.run.vm09.stdout:(102/152): python3-pyparsing-2.4.7-9.el9.0.1.no 4.5 MB/s | 150 kB 00:00 2026-04-01T13:53:30.054 INFO:teuthology.orchestra.run.vm08.stdout:(96/152): nvme-cli-2.13-1.el9.x86_64.rpm 23 MB/s | 1.0 MB 00:00 2026-04-01T13:53:30.056 INFO:teuthology.orchestra.run.vm08.stdout:(97/152): pciutils-3.7.0-7.el9.x86_64.rpm 3.1 MB/s | 92 kB 00:00 2026-04-01T13:53:30.063 INFO:teuthology.orchestra.run.vm09.stdout:(103/152): python3-requests-2.25.1-10.el9_6.noa 3.9 MB/s | 115 kB 00:00 2026-04-01T13:53:30.066 INFO:teuthology.orchestra.run.vm09.stdout:(104/152): nvme-cli-2.13-1.el9.x86_64.rpm 6.1 MB/s | 1.0 MB 00:00 2026-04-01T13:53:30.082 INFO:teuthology.orchestra.run.vm09.stdout:(105/152): python3-urllib3-1.26.5-6.el9_7.1.noa 6.0 MB/s | 191 kB 00:00 2026-04-01T13:53:30.093 INFO:teuthology.orchestra.run.vm08.stdout:(98/152): python3-cffi-1.14.5-5.el9.x86_64.rpm 6.1 MB/s | 241 kB 00:00 2026-04-01T13:53:30.102 INFO:teuthology.orchestra.run.vm09.stdout:(106/152): smartmontools-7.2-9.el9.x86_64.rpm 14 MB/s | 551 kB 00:00 2026-04-01T13:53:30.113 INFO:teuthology.orchestra.run.vm09.stdout:(107/152): zip-3.0-35.el9.x86_64.rpm 8.3 MB/s | 263 kB 00:00 2026-04-01T13:53:30.122 INFO:teuthology.orchestra.run.vm08.stdout:(99/152): python3-ply-3.11-14.el9.0.1.noarch.rp 3.5 MB/s | 103 kB 00:00 2026-04-01T13:53:30.129 INFO:teuthology.orchestra.run.vm09.stdout:(108/152): boost-program-options-1.75.0-13.el9_ 3.7 MB/s | 104 kB 00:00 2026-04-01T13:53:30.136 INFO:teuthology.orchestra.run.vm06.stdout:(41/152): libarrow-9.0.0-15.el9.x86_64.rpm 50 MB/s | 4.4 MB 00:00 2026-04-01T13:53:30.138 INFO:teuthology.orchestra.run.vm06.stdout:(42/152): libarrow-doc-9.0.0-15.el9.noarch.rpm 10 MB/s | 25 kB 00:00 2026-04-01T13:53:30.139 INFO:teuthology.orchestra.run.vm09.stdout:(109/152): flexiblas-3.0.4-8.el9.0.1.x86_64.rpm 1.1 MB/s | 30 kB 00:00 2026-04-01T13:53:30.141 INFO:teuthology.orchestra.run.vm06.stdout:(43/152): liboath-2.6.12-1.el9.x86_64.rpm 20 MB/s | 49 kB 00:00 2026-04-01T13:53:30.144 INFO:teuthology.orchestra.run.vm06.stdout:(44/152): libunwind-1.6.2-1.el9.x86_64.rpm 25 MB/s | 67 kB 00:00 2026-04-01T13:53:30.149 INFO:teuthology.orchestra.run.vm06.stdout:(45/152): luarocks-3.9.2-5.el9.noarch.rpm 31 MB/s | 151 kB 00:00 2026-04-01T13:53:30.150 INFO:teuthology.orchestra.run.vm08.stdout:(100/152): python3-pycparser-2.20-6.el9.noarch. 4.4 MB/s | 124 kB 00:00 2026-04-01T13:53:30.166 INFO:teuthology.orchestra.run.vm06.stdout:(46/152): parquet-libs-9.0.0-15.el9.x86_64.rpm 48 MB/s | 838 kB 00:00 2026-04-01T13:53:30.171 INFO:teuthology.orchestra.run.vm09.stdout:(110/152): flexiblas-openblas-openmp-3.0.4-8.el 475 kB/s | 15 kB 00:00 2026-04-01T13:53:30.180 INFO:teuthology.orchestra.run.vm08.stdout:(101/152): python3-pyparsing-2.4.7-9.el9.0.1.no 5.0 MB/s | 150 kB 00:00 2026-04-01T13:53:30.207 INFO:teuthology.orchestra.run.vm09.stdout:(111/152): unzip-6.0-59.el9.x86_64.rpm 1.2 MB/s | 180 kB 00:00 2026-04-01T13:53:30.208 INFO:teuthology.orchestra.run.vm06.stdout:(47/152): ceph-mgr-dashboard-20.2.0-9.ge2ce8426 12 MB/s | 15 MB 00:01 2026-04-01T13:53:30.208 INFO:teuthology.orchestra.run.vm08.stdout:(102/152): python3-requests-2.25.1-10.el9_6.noa 4.1 MB/s | 115 kB 00:00 2026-04-01T13:53:30.209 INFO:teuthology.orchestra.run.vm09.stdout:(112/152): libnbd-1.20.3-4.el9.x86_64.rpm 4.4 MB/s | 171 kB 00:00 2026-04-01T13:53:30.213 INFO:teuthology.orchestra.run.vm06.stdout:(48/152): python3-asyncssh-2.13.2-5.el9.noarch. 11 MB/s | 548 kB 00:00 2026-04-01T13:53:30.214 INFO:teuthology.orchestra.run.vm06.stdout:(49/152): python3-autocommand-2.2.2-8.el9.noarc 4.5 MB/s | 29 kB 00:00 2026-04-01T13:53:30.217 INFO:teuthology.orchestra.run.vm06.stdout:(50/152): python3-backports-tarfile-1.2.0-1.el9 20 MB/s | 60 kB 00:00 2026-04-01T13:53:30.220 INFO:teuthology.orchestra.run.vm06.stdout:(51/152): python3-bcrypt-3.2.2-1.el9.x86_64.rpm 7.8 MB/s | 43 kB 00:00 2026-04-01T13:53:30.220 INFO:teuthology.orchestra.run.vm09.stdout:(113/152): flexiblas-netlib-3.0.4-8.el9.0.1.x86 33 MB/s | 3.0 MB 00:00 2026-04-01T13:53:30.221 INFO:teuthology.orchestra.run.vm06.stdout:(52/152): python3-cachetools-4.2.4-1.el9.noarch 7.0 MB/s | 32 kB 00:00 2026-04-01T13:53:30.222 INFO:teuthology.orchestra.run.vm06.stdout:(53/152): python3-certifi-2023.05.07-4.el9.noar 6.3 MB/s | 14 kB 00:00 2026-04-01T13:53:30.227 INFO:teuthology.orchestra.run.vm06.stdout:(54/152): python3-cheroot-10.0.1-5.el9.noarch.r 31 MB/s | 173 kB 00:00 2026-04-01T13:53:30.230 INFO:teuthology.orchestra.run.vm06.stdout:(55/152): python3-cherrypy-18.10.0-5.el9.noarch 38 MB/s | 290 kB 00:00 2026-04-01T13:53:30.234 INFO:teuthology.orchestra.run.vm06.stdout:(56/152): python3-google-auth-2.45.0-1.el9.noar 37 MB/s | 254 kB 00:00 2026-04-01T13:53:30.241 INFO:teuthology.orchestra.run.vm09.stdout:(114/152): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.4 MB/s | 44 kB 00:00 2026-04-01T13:53:30.242 INFO:teuthology.orchestra.run.vm08.stdout:(103/152): python3-urllib3-1.26.5-6.el9_7.1.noa 5.4 MB/s | 191 kB 00:00 2026-04-01T13:53:30.242 INFO:teuthology.orchestra.run.vm09.stdout:(115/152): libpmemobj-1.12.1-1.el9.x86_64.rpm 4.5 MB/s | 159 kB 00:00 2026-04-01T13:53:30.243 INFO:teuthology.orchestra.run.vm06.stdout:(57/152): python3-grpcio-tools-1.46.7-10.el9.x8 16 MB/s | 144 kB 00:00 2026-04-01T13:53:30.251 INFO:teuthology.orchestra.run.vm06.stdout:(58/152): python3-influxdb-5.3.1-1.el9.noarch.r 18 MB/s | 139 kB 00:00 2026-04-01T13:53:30.274 INFO:teuthology.orchestra.run.vm09.stdout:(116/152): librdkafka-1.6.1-102.el9.x86_64.rpm 12 MB/s | 662 kB 00:00 2026-04-01T13:53:30.308 INFO:teuthology.orchestra.run.vm08.stdout:(104/152): smartmontools-7.2-9.el9.x86_64.rpm 8.2 MB/s | 551 kB 00:00 2026-04-01T13:53:30.308 INFO:teuthology.orchestra.run.vm09.stdout:(117/152): libxslt-1.1.34-13.el9_6.x86_64.rpm 3.5 MB/s | 239 kB 00:00 2026-04-01T13:53:30.309 INFO:teuthology.orchestra.run.vm06.stdout:(59/152): python3-isodate-0.6.1-3.el9.noarch.rp 993 kB/s | 56 kB 00:00 2026-04-01T13:53:30.310 INFO:teuthology.orchestra.run.vm09.stdout:(118/152): libstoragemgmt-1.10.1-1.el9.x86_64.r 3.5 MB/s | 243 kB 00:00 2026-04-01T13:53:30.312 INFO:teuthology.orchestra.run.vm09.stdout:(119/152): lttng-ust-2.12.0-6.el9.x86_64.rpm 7.5 MB/s | 282 kB 00:00 2026-04-01T13:53:30.312 INFO:teuthology.orchestra.run.vm06.stdout:(60/152): python3-jaraco-8.2.1-3.el9.noarch.rpm 2.8 MB/s | 11 kB 00:00 2026-04-01T13:53:30.318 INFO:teuthology.orchestra.run.vm06.stdout:(61/152): python3-grpcio-1.46.7-10.el9.x86_64.r 23 MB/s | 2.0 MB 00:00 2026-04-01T13:53:30.319 INFO:teuthology.orchestra.run.vm06.stdout:(62/152): python3-jaraco-classes-3.2.1-5.el9.no 2.3 MB/s | 18 kB 00:00 2026-04-01T13:53:30.321 INFO:teuthology.orchestra.run.vm06.stdout:(63/152): python3-jaraco-collections-3.0.0-8.el 8.4 MB/s | 23 kB 00:00 2026-04-01T13:53:30.322 INFO:teuthology.orchestra.run.vm06.stdout:(64/152): python3-jaraco-context-6.0.1-3.el9.no 7.2 MB/s | 20 kB 00:00 2026-04-01T13:53:30.324 INFO:teuthology.orchestra.run.vm06.stdout:(65/152): python3-jaraco-functools-3.5.0-2.el9. 7.8 MB/s | 19 kB 00:00 2026-04-01T13:53:30.325 INFO:teuthology.orchestra.run.vm06.stdout:(66/152): python3-jaraco-text-4.0.0-2.el9.noarc 11 MB/s | 26 kB 00:00 2026-04-01T13:53:30.330 INFO:teuthology.orchestra.run.vm06.stdout:(67/152): python3-more-itertools-8.12.0-2.el9.n 15 MB/s | 79 kB 00:00 2026-04-01T13:53:30.335 INFO:teuthology.orchestra.run.vm06.stdout:(68/152): python3-msgpack-1.0.3-2.el9.x86_64.rp 18 MB/s | 86 kB 00:00 2026-04-01T13:53:30.341 INFO:teuthology.orchestra.run.vm08.stdout:(105/152): unzip-6.0-59.el9.x86_64.rpm 5.4 MB/s | 180 kB 00:00 2026-04-01T13:53:30.345 INFO:teuthology.orchestra.run.vm06.stdout:(69/152): python3-kubernetes-26.1.0-3.el9.noarc 50 MB/s | 1.0 MB 00:00 2026-04-01T13:53:30.346 INFO:teuthology.orchestra.run.vm06.stdout:(70/152): python3-natsort-7.1.1-5.el9.noarch.rp 5.3 MB/s | 58 kB 00:00 2026-04-01T13:53:30.352 INFO:teuthology.orchestra.run.vm06.stdout:(71/152): python3-portend-3.1.0-2.el9.noarch.rp 2.3 MB/s | 16 kB 00:00 2026-04-01T13:53:30.355 INFO:teuthology.orchestra.run.vm06.stdout:(72/152): python3-pyOpenSSL-21.0.0-1.el9.noarch 9.9 MB/s | 90 kB 00:00 2026-04-01T13:53:30.359 INFO:teuthology.orchestra.run.vm06.stdout:(73/152): python3-repoze-lru-0.7-16.el9.noarch. 4.6 MB/s | 31 kB 00:00 2026-04-01T13:53:30.360 INFO:teuthology.orchestra.run.vm09.stdout:(120/152): openblas-0.3.29-1.el9.x86_64.rpm 819 kB/s | 41 kB 00:00 2026-04-01T13:53:30.362 INFO:teuthology.orchestra.run.vm09.stdout:(121/152): lua-5.4.4-4.el9.x86_64.rpm 3.4 MB/s | 187 kB 00:00 2026-04-01T13:53:30.364 INFO:teuthology.orchestra.run.vm06.stdout:(74/152): python3-rsa-4.9-2.el9.noarch.rpm 11 MB/s | 59 kB 00:00 2026-04-01T13:53:30.366 INFO:teuthology.orchestra.run.vm06.stdout:(75/152): python3-routes-2.5.1-5.el9.noarch.rpm 17 MB/s | 188 kB 00:00 2026-04-01T13:53:30.368 INFO:teuthology.orchestra.run.vm06.stdout:(76/152): python3-saml-1.16.0-1.el9.noarch.rpm 32 MB/s | 125 kB 00:00 2026-04-01T13:53:30.371 INFO:teuthology.orchestra.run.vm06.stdout:(77/152): python3-tempora-5.0.0-2.el9.noarch.rp 7.8 MB/s | 36 kB 00:00 2026-04-01T13:53:30.371 INFO:teuthology.orchestra.run.vm08.stdout:(106/152): python3-cryptography-36.0.1-5.el9_6. 3.7 MB/s | 1.2 MB 00:00 2026-04-01T13:53:30.373 INFO:teuthology.orchestra.run.vm06.stdout:(78/152): python3-typing-extensions-4.15.0-1.el 18 MB/s | 86 kB 00:00 2026-04-01T13:53:30.373 INFO:teuthology.orchestra.run.vm08.stdout:(107/152): zip-3.0-35.el9.x86_64.rpm 7.9 MB/s | 263 kB 00:00 2026-04-01T13:53:30.376 INFO:teuthology.orchestra.run.vm06.stdout:(79/152): python3-websocket-client-1.2.3-2.el9. 17 MB/s | 90 kB 00:00 2026-04-01T13:53:30.379 INFO:teuthology.orchestra.run.vm06.stdout:(80/152): python3-xmlsec-1.3.13-1.el9.x86_64.rp 9.1 MB/s | 48 kB 00:00 2026-04-01T13:53:30.382 INFO:teuthology.orchestra.run.vm06.stdout:(81/152): python3-xmltodict-0.12.0-15.el9.noarc 3.8 MB/s | 22 kB 00:00 2026-04-01T13:53:30.384 INFO:teuthology.orchestra.run.vm06.stdout:(82/152): python3-zc-lockfile-2.0-10.el9.noarch 3.3 MB/s | 20 kB 00:00 2026-04-01T13:53:30.390 INFO:teuthology.orchestra.run.vm09.stdout:(122/152): perl-Benchmark-1.23-481.1.el9_6.noar 861 kB/s | 25 kB 00:00 2026-04-01T13:53:30.393 INFO:teuthology.orchestra.run.vm06.stdout:(83/152): s3cmd-2.4.0-1.el9.noarch.rpm 26 MB/s | 206 kB 00:00 2026-04-01T13:53:30.395 INFO:teuthology.orchestra.run.vm06.stdout:(84/152): re2-20211101-20.el9.x86_64.rpm 14 MB/s | 191 kB 00:00 2026-04-01T13:53:30.396 INFO:teuthology.orchestra.run.vm09.stdout:(123/152): perl-Test-Harness-3.42-461.el9.noarc 7.6 MB/s | 267 kB 00:00 2026-04-01T13:53:30.404 INFO:teuthology.orchestra.run.vm08.stdout:(108/152): flexiblas-3.0.4-8.el9.0.1.x86_64.rpm 978 kB/s | 30 kB 00:00 2026-04-01T13:53:30.407 INFO:teuthology.orchestra.run.vm08.stdout:(109/152): boost-program-options-1.75.0-13.el9_ 2.9 MB/s | 104 kB 00:00 2026-04-01T13:53:30.418 INFO:teuthology.orchestra.run.vm06.stdout:(85/152): thrift-0.15.0-4.el9.x86_64.rpm 63 MB/s | 1.6 MB 00:00 2026-04-01T13:53:30.436 INFO:teuthology.orchestra.run.vm08.stdout:(110/152): flexiblas-openblas-openmp-3.0.4-8.el 515 kB/s | 15 kB 00:00 2026-04-01T13:53:30.465 INFO:teuthology.orchestra.run.vm09.stdout:(124/152): protobuf-3.14.0-17.el9_7.x86_64.rpm 14 MB/s | 1.0 MB 00:00 2026-04-01T13:53:30.489 INFO:teuthology.orchestra.run.vm09.stdout:(125/152): openblas-openmp-0.3.29-1.el9.x86_64. 30 MB/s | 5.3 MB 00:00 2026-04-01T13:53:30.505 INFO:teuthology.orchestra.run.vm09.stdout:(126/152): python3-devel-3.9.23-2.el9.x86_64.rp 5.0 MB/s | 205 kB 00:00 2026-04-01T13:53:30.519 INFO:teuthology.orchestra.run.vm09.stdout:(127/152): python3-jinja2-2.11.3-8.el9_5.noarch 7.3 MB/s | 228 kB 00:00 2026-04-01T13:53:30.532 INFO:teuthology.orchestra.run.vm09.stdout:(128/152): python3-jmespath-1.0.1-1.el9_7.noarc 1.6 MB/s | 43 kB 00:00 2026-04-01T13:53:30.549 INFO:teuthology.orchestra.run.vm09.stdout:(129/152): python3-libstoragemgmt-1.10.1-1.el9. 5.5 MB/s | 166 kB 00:00 2026-04-01T13:53:30.574 INFO:teuthology.orchestra.run.vm06.stdout:(86/152): bzip2-1.0.8-10.el9_5.x86_64.rpm 288 kB/s | 51 kB 00:00 2026-04-01T13:53:30.576 INFO:teuthology.orchestra.run.vm09.stdout:(130/152): python3-markupsafe-1.1.1-12.el9.x86_ 1.2 MB/s | 32 kB 00:00 2026-04-01T13:53:30.633 INFO:teuthology.orchestra.run.vm06.stdout:(87/152): c-ares-1.19.1-2.el9_4.x86_64.rpm 510 kB/s | 110 kB 00:00 2026-04-01T13:53:30.653 INFO:teuthology.orchestra.run.vm09.stdout:(131/152): python3-babel-2.9.1-2.el9.noarch.rpm 23 MB/s | 5.8 MB 00:00 2026-04-01T13:53:30.678 INFO:teuthology.orchestra.run.vm06.stdout:(88/152): fuse-2.9.9-17.el9.x86_64.rpm 1.7 MB/s | 78 kB 00:00 2026-04-01T13:53:30.678 INFO:teuthology.orchestra.run.vm08.stdout:(111/152): ceph-test-20.2.0-9.ge2ce8426bc5.el9. 15 MB/s | 85 MB 00:05 2026-04-01T13:53:30.678 INFO:teuthology.orchestra.run.vm09.stdout:(132/152): python3-lxml-4.6.5-3.el9.x86_64.rpm 8.2 MB/s | 1.2 MB 00:00 2026-04-01T13:53:30.680 INFO:teuthology.orchestra.run.vm08.stdout:(112/152): libnbd-1.20.3-4.el9.x86_64.rpm 700 kB/s | 171 kB 00:00 2026-04-01T13:53:30.689 INFO:teuthology.orchestra.run.vm08.stdout:(113/152): flexiblas-netlib-3.0.4-8.el9.0.1.x86 10 MB/s | 3.0 MB 00:00 2026-04-01T13:53:30.689 INFO:teuthology.orchestra.run.vm09.stdout:(133/152): python3-numpy-f2py-1.23.5-2.el9_7.x8 10 MB/s | 368 kB 00:00 2026-04-01T13:53:30.693 INFO:teuthology.orchestra.run.vm06.stdout:(89/152): cryptsetup-2.7.2-4.el9.x86_64.rpm 2.6 MB/s | 310 kB 00:00 2026-04-01T13:53:30.705 INFO:teuthology.orchestra.run.vm06.stdout:(90/152): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-01T13:53:30.708 INFO:teuthology.orchestra.run.vm09.stdout:(134/152): python3-packaging-20.9-5.el9.noarch. 2.3 MB/s | 69 kB 00:00 2026-04-01T13:53:30.712 INFO:teuthology.orchestra.run.vm08.stdout:(114/152): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.3 MB/s | 44 kB 00:00 2026-04-01T13:53:30.724 INFO:teuthology.orchestra.run.vm09.stdout:(135/152): python3-protobuf-3.14.0-17.el9_7.noa 6.8 MB/s | 237 kB 00:00 2026-04-01T13:53:30.724 INFO:teuthology.orchestra.run.vm06.stdout:(91/152): libconfig-1.7.2-9.el9.x86_64.rpm 2.2 MB/s | 71 kB 00:00 2026-04-01T13:53:30.731 INFO:teuthology.orchestra.run.vm08.stdout:(115/152): librdkafka-1.6.1-102.el9.x86_64.rpm 16 MB/s | 662 kB 00:00 2026-04-01T13:53:30.737 INFO:teuthology.orchestra.run.vm09.stdout:(136/152): python3-pyasn1-0.4.8-7.el9_7.noarch. 4.5 MB/s | 132 kB 00:00 2026-04-01T13:53:30.764 INFO:teuthology.orchestra.run.vm08.stdout:(116/152): libxslt-1.1.34-13.el9_6.x86_64.rpm 7.1 MB/s | 239 kB 00:00 2026-04-01T13:53:30.769 INFO:teuthology.orchestra.run.vm09.stdout:(137/152): python3-requests-oauthlib-1.3.0-12.e 1.3 MB/s | 43 kB 00:00 2026-04-01T13:53:30.772 INFO:teuthology.orchestra.run.vm09.stdout:(138/152): python3-pyasn1-modules-0.4.8-7.el9_7 4.3 MB/s | 210 kB 00:00 2026-04-01T13:53:30.774 INFO:teuthology.orchestra.run.vm06.stdout:(92/152): libquadmath-11.5.0-11.el9.x86_64.rpm 3.6 MB/s | 184 kB 00:00 2026-04-01T13:53:30.775 INFO:teuthology.orchestra.run.vm08.stdout:(117/152): libstoragemgmt-1.10.1-1.el9.x86_64.r 3.8 MB/s | 243 kB 00:00 2026-04-01T13:53:30.787 INFO:teuthology.orchestra.run.vm09.stdout:(139/152): python3-numpy-1.23.5-2.el9_7.x86_64. 27 MB/s | 5.8 MB 00:00 2026-04-01T13:53:30.787 INFO:teuthology.orchestra.run.vm06.stdout:(93/152): libgfortran-11.5.0-11.el9.x86_64.rpm 9.4 MB/s | 794 kB 00:00 2026-04-01T13:53:30.794 INFO:teuthology.orchestra.run.vm08.stdout:(118/152): lttng-ust-2.12.0-6.el9.x86_64.rpm 9.4 MB/s | 282 kB 00:00 2026-04-01T13:53:30.799 INFO:teuthology.orchestra.run.vm09.stdout:(140/152): python3-toml-0.10.2-6.el9.0.1.noarch 1.6 MB/s | 44 kB 00:00 2026-04-01T13:53:30.803 INFO:teuthology.orchestra.run.vm06.stdout:(94/152): lmdb-libs-0.9.29-3.el9.x86_64.rpm 2.1 MB/s | 60 kB 00:00 2026-04-01T13:53:30.814 INFO:teuthology.orchestra.run.vm06.stdout:(95/152): mailcap-2.1.49-5.el9.0.2.noarch.rpm 1.2 MB/s | 32 kB 00:00 2026-04-01T13:53:30.816 INFO:teuthology.orchestra.run.vm09.stdout:(141/152): qatlib-24.09.0-1.el9.x86_64.rpm 7.5 MB/s | 221 kB 00:00 2026-04-01T13:53:30.821 INFO:teuthology.orchestra.run.vm08.stdout:(119/152): openblas-0.3.29-1.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-01T13:53:30.830 INFO:teuthology.orchestra.run.vm08.stdout:(120/152): lua-5.4.4-4.el9.x86_64.rpm 3.4 MB/s | 187 kB 00:00 2026-04-01T13:53:30.838 INFO:teuthology.orchestra.run.vm09.stdout:(142/152): qatlib-service-24.09.0-1.el9.x86_64. 936 kB/s | 36 kB 00:00 2026-04-01T13:53:30.842 INFO:teuthology.orchestra.run.vm06.stdout:(96/152): pciutils-3.7.0-7.el9.x86_64.rpm 3.2 MB/s | 92 kB 00:00 2026-04-01T13:53:30.843 INFO:teuthology.orchestra.run.vm09.stdout:(143/152): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.3 MB/s | 65 kB 00:00 2026-04-01T13:53:30.844 INFO:teuthology.orchestra.run.vm08.stdout:(121/152): libpmemobj-1.12.1-1.el9.x86_64.rpm 965 kB/s | 159 kB 00:00 2026-04-01T13:53:30.859 INFO:teuthology.orchestra.run.vm08.stdout:(122/152): perl-Benchmark-1.23-481.1.el9_6.noar 881 kB/s | 25 kB 00:00 2026-04-01T13:53:30.879 INFO:teuthology.orchestra.run.vm09.stdout:(144/152): xmlsec1-1.2.29-13.el9.x86_64.rpm 5.2 MB/s | 188 kB 00:00 2026-04-01T13:53:30.881 INFO:teuthology.orchestra.run.vm09.stdout:(145/152): socat-1.7.4.1-8.el9.x86_64.rpm 6.8 MB/s | 299 kB 00:00 2026-04-01T13:53:30.886 INFO:teuthology.orchestra.run.vm06.stdout:(97/152): python3-cffi-1.14.5-5.el9.x86_64.rpm 5.5 MB/s | 241 kB 00:00 2026-04-01T13:53:30.894 INFO:teuthology.orchestra.run.vm06.stdout:(98/152): nvme-cli-2.13-1.el9.x86_64.rpm 11 MB/s | 1.0 MB 00:00 2026-04-01T13:53:30.899 INFO:teuthology.orchestra.run.vm08.stdout:(123/152): perl-Test-Harness-3.42-461.el9.noarc 4.7 MB/s | 267 kB 00:00 2026-04-01T13:53:30.908 INFO:teuthology.orchestra.run.vm09.stdout:(146/152): xmlsec1-openssl-1.2.29-13.el9.x86_64 3.1 MB/s | 89 kB 00:00 2026-04-01T13:53:30.909 INFO:teuthology.orchestra.run.vm09.stdout:(147/152): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.2 MB/s | 63 kB 00:00 2026-04-01T13:53:30.924 INFO:teuthology.orchestra.run.vm06.stdout:(99/152): python3-ply-3.11-14.el9.0.1.noarch.rp 3.3 MB/s | 103 kB 00:00 2026-04-01T13:53:30.939 INFO:teuthology.orchestra.run.vm09.stdout:(148/152): lua-devel-5.4.4-4.el9.x86_64.rpm 688 kB/s | 21 kB 00:00 2026-04-01T13:53:30.964 INFO:teuthology.orchestra.run.vm06.stdout:(100/152): python3-pycparser-2.20-6.el9.noarch. 3.1 MB/s | 124 kB 00:00 2026-04-01T13:53:30.972 INFO:teuthology.orchestra.run.vm06.stdout:(101/152): python3-cryptography-36.0.1-5.el9_6. 14 MB/s | 1.2 MB 00:00 2026-04-01T13:53:30.982 INFO:teuthology.orchestra.run.vm09.stdout:(149/152): protobuf-compiler-3.14.0-17.el9_7.x8 12 MB/s | 862 kB 00:00 2026-04-01T13:53:30.983 INFO:teuthology.orchestra.run.vm08.stdout:(124/152): openblas-openmp-0.3.29-1.el9.x86_64. 33 MB/s | 5.3 MB 00:00 2026-04-01T13:53:30.994 INFO:teuthology.orchestra.run.vm06.stdout:(102/152): python3-pyparsing-2.4.7-9.el9.0.1.no 5.0 MB/s | 150 kB 00:00 2026-04-01T13:53:31.001 INFO:teuthology.orchestra.run.vm06.stdout:(103/152): python3-requests-2.25.1-10.el9_6.noa 3.9 MB/s | 115 kB 00:00 2026-04-01T13:53:31.015 INFO:teuthology.orchestra.run.vm08.stdout:(125/152): python3-devel-3.9.23-2.el9.x86_64.rp 6.2 MB/s | 205 kB 00:00 2026-04-01T13:53:31.034 INFO:teuthology.orchestra.run.vm06.stdout:(104/152): smartmontools-7.2-9.el9.x86_64.rpm 16 MB/s | 551 kB 00:00 2026-04-01T13:53:31.049 INFO:teuthology.orchestra.run.vm06.stdout:(105/152): python3-urllib3-1.26.5-6.el9_7.1.noa 3.4 MB/s | 191 kB 00:00 2026-04-01T13:53:31.051 INFO:teuthology.orchestra.run.vm08.stdout:(126/152): python3-jinja2-2.11.3-8.el9_5.noarch 6.3 MB/s | 228 kB 00:00 2026-04-01T13:53:31.067 INFO:teuthology.orchestra.run.vm06.stdout:(106/152): unzip-6.0-59.el9.x86_64.rpm 5.4 MB/s | 180 kB 00:00 2026-04-01T13:53:31.079 INFO:teuthology.orchestra.run.vm08.stdout:(127/152): python3-jmespath-1.0.1-1.el9_7.noarc 1.5 MB/s | 43 kB 00:00 2026-04-01T13:53:31.081 INFO:teuthology.orchestra.run.vm06.stdout:(107/152): zip-3.0-35.el9.x86_64.rpm 8.0 MB/s | 263 kB 00:00 2026-04-01T13:53:31.098 INFO:teuthology.orchestra.run.vm06.stdout:(108/152): boost-program-options-1.75.0-13.el9_ 3.4 MB/s | 104 kB 00:00 2026-04-01T13:53:31.099 INFO:teuthology.orchestra.run.vm08.stdout:(128/152): protobuf-3.14.0-17.el9_7.x86_64.rpm 4.2 MB/s | 1.0 MB 00:00 2026-04-01T13:53:31.108 INFO:teuthology.orchestra.run.vm06.stdout:(109/152): flexiblas-3.0.4-8.el9.0.1.x86_64.rpm 1.1 MB/s | 30 kB 00:00 2026-04-01T13:53:31.108 INFO:teuthology.orchestra.run.vm08.stdout:(129/152): python3-libstoragemgmt-1.10.1-1.el9. 5.5 MB/s | 166 kB 00:00 2026-04-01T13:53:31.123 INFO:teuthology.orchestra.run.vm08.stdout:(130/152): python3-babel-2.9.1-2.el9.noarch.rpm 26 MB/s | 5.8 MB 00:00 2026-04-01T13:53:31.135 INFO:teuthology.orchestra.run.vm06.stdout:(110/152): flexiblas-openblas-openmp-3.0.4-8.el 559 kB/s | 15 kB 00:00 2026-04-01T13:53:31.135 INFO:teuthology.orchestra.run.vm08.stdout:(131/152): python3-markupsafe-1.1.1-12.el9.x86_ 1.2 MB/s | 32 kB 00:00 2026-04-01T13:53:31.172 INFO:teuthology.orchestra.run.vm06.stdout:(111/152): libnbd-1.20.3-4.el9.x86_64.rpm 4.5 MB/s | 171 kB 00:00 2026-04-01T13:53:31.176 INFO:teuthology.orchestra.run.vm08.stdout:(132/152): python3-numpy-f2py-1.23.5-2.el9_7.x8 8.9 MB/s | 368 kB 00:00 2026-04-01T13:53:31.228 INFO:teuthology.orchestra.run.vm08.stdout:(133/152): python3-packaging-20.9-5.el9.noarch. 1.3 MB/s | 69 kB 00:00 2026-04-01T13:53:31.320 INFO:teuthology.orchestra.run.vm06.stdout:(112/152): libpmemobj-1.12.1-1.el9.x86_64.rpm 1.1 MB/s | 159 kB 00:00 2026-04-01T13:53:31.325 INFO:teuthology.orchestra.run.vm08.stdout:(134/152): python3-protobuf-3.14.0-17.el9_7.noa 2.4 MB/s | 237 kB 00:00 2026-04-01T13:53:31.360 INFO:teuthology.orchestra.run.vm06.stdout:(113/152): flexiblas-netlib-3.0.4-8.el9.0.1.x86 11 MB/s | 3.0 MB 00:00 2026-04-01T13:53:31.360 INFO:teuthology.orchestra.run.vm09.stdout:(150/152): librbd1-20.2.0-9.ge2ce8426bc5.el9.cl 7.5 MB/s | 2.8 MB 00:00 2026-04-01T13:53:31.404 INFO:teuthology.orchestra.run.vm06.stdout:(114/152): librabbitmq-0.11.0-7.el9.x86_64.rpm 528 kB/s | 44 kB 00:00 2026-04-01T13:53:31.437 INFO:teuthology.orchestra.run.vm08.stdout:(135/152): python3-numpy-1.23.5-2.el9_7.x86_64. 18 MB/s | 5.8 MB 00:00 2026-04-01T13:53:31.462 INFO:teuthology.orchestra.run.vm06.stdout:(115/152): librdkafka-1.6.1-102.el9.x86_64.rpm 6.4 MB/s | 662 kB 00:00 2026-04-01T13:53:31.485 INFO:teuthology.orchestra.run.vm08.stdout:(136/152): python3-pyasn1-0.4.8-7.el9_7.noarch. 824 kB/s | 132 kB 00:00 2026-04-01T13:53:31.505 INFO:teuthology.orchestra.run.vm06.stdout:(116/152): libstoragemgmt-1.10.1-1.el9.x86_64.r 2.4 MB/s | 243 kB 00:00 2026-04-01T13:53:31.538 INFO:teuthology.orchestra.run.vm08.stdout:(137/152): python3-lxml-4.6.5-3.el9.x86_64.rpm 2.7 MB/s | 1.2 MB 00:00 2026-04-01T13:53:31.556 INFO:teuthology.orchestra.run.vm06.stdout:(117/152): libxslt-1.1.34-13.el9_6.x86_64.rpm 2.5 MB/s | 239 kB 00:00 2026-04-01T13:53:31.574 INFO:teuthology.orchestra.run.vm08.stdout:(138/152): python3-requests-oauthlib-1.3.0-12.e 482 kB/s | 43 kB 00:00 2026-04-01T13:53:31.629 INFO:teuthology.orchestra.run.vm06.stdout:(118/152): lttng-ust-2.12.0-6.el9.x86_64.rpm 2.2 MB/s | 282 kB 00:00 2026-04-01T13:53:31.629 INFO:teuthology.orchestra.run.vm08.stdout:(139/152): python3-pyasn1-modules-0.4.8-7.el9_7 1.1 MB/s | 210 kB 00:00 2026-04-01T13:53:31.661 INFO:teuthology.orchestra.run.vm06.stdout:(119/152): lua-5.4.4-4.el9.x86_64.rpm 1.7 MB/s | 187 kB 00:00 2026-04-01T13:53:31.662 INFO:teuthology.orchestra.run.vm08.stdout:(140/152): python3-toml-0.10.2-6.el9.0.1.noarch 504 kB/s | 44 kB 00:00 2026-04-01T13:53:31.719 INFO:teuthology.orchestra.run.vm08.stdout:(141/152): qatlib-24.09.0-1.el9.x86_64.rpm 2.4 MB/s | 221 kB 00:00 2026-04-01T13:53:31.720 INFO:teuthology.orchestra.run.vm06.stdout:(120/152): openblas-0.3.29-1.el9.x86_64.rpm 456 kB/s | 41 kB 00:00 2026-04-01T13:53:31.810 INFO:teuthology.orchestra.run.vm09.stdout:(151/152): python3-scipy-1.9.3-2.el9.x86_64.rpm 18 MB/s | 19 MB 00:01 2026-04-01T13:53:31.826 INFO:teuthology.orchestra.run.vm06.stdout:(121/152): perl-Benchmark-1.23-481.1.el9_6.noar 239 kB/s | 25 kB 00:00 2026-04-01T13:53:31.827 INFO:teuthology.orchestra.run.vm08.stdout:(142/152): qatlib-service-24.09.0-1.el9.x86_64. 221 kB/s | 36 kB 00:00 2026-04-01T13:53:31.934 INFO:teuthology.orchestra.run.vm08.stdout:(143/152): qatzip-libs-1.3.1-1.el9.x86_64.rpm 304 kB/s | 65 kB 00:00 2026-04-01T13:53:31.971 INFO:teuthology.orchestra.run.vm06.stdout:(122/152): perl-Test-Harness-3.42-461.el9.noarc 1.8 MB/s | 267 kB 00:00 2026-04-01T13:53:31.973 INFO:teuthology.orchestra.run.vm09.stdout:(152/152): librados2-20.2.0-9.ge2ce8426bc5.el9. 3.4 MB/s | 3.5 MB 00:01 2026-04-01T13:53:31.976 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-04-01T13:53:31.976 INFO:teuthology.orchestra.run.vm09.stdout:Total 28 MB/s | 274 MB 00:09 2026-04-01T13:53:31.984 INFO:teuthology.orchestra.run.vm08.stdout:(144/152): socat-1.7.4.1-8.el9.x86_64.rpm 1.9 MB/s | 299 kB 00:00 2026-04-01T13:53:31.990 INFO:teuthology.orchestra.run.vm08.stdout:(145/152): xmlsec1-1.2.29-13.el9.x86_64.rpm 3.3 MB/s | 188 kB 00:00 2026-04-01T13:53:31.998 INFO:teuthology.orchestra.run.vm06.stdout:(123/152): openblas-openmp-0.3.29-1.el9.x86_64. 16 MB/s | 5.3 MB 00:00 2026-04-01T13:53:32.018 INFO:teuthology.orchestra.run.vm08.stdout:(146/152): xmlsec1-openssl-1.2.29-13.el9.x86_64 2.6 MB/s | 89 kB 00:00 2026-04-01T13:53:32.018 INFO:teuthology.orchestra.run.vm06.stdout:(124/152): protobuf-3.14.0-17.el9_7.x86_64.rpm 21 MB/s | 1.0 MB 00:00 2026-04-01T13:53:32.021 INFO:teuthology.orchestra.run.vm08.stdout:(147/152): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.0 MB/s | 63 kB 00:00 2026-04-01T13:53:32.045 INFO:teuthology.orchestra.run.vm08.stdout:(148/152): lua-devel-5.4.4-4.el9.x86_64.rpm 803 kB/s | 21 kB 00:00 2026-04-01T13:53:32.047 INFO:teuthology.orchestra.run.vm06.stdout:(125/152): python3-devel-3.9.23-2.el9.x86_64.rp 7.1 MB/s | 205 kB 00:00 2026-04-01T13:53:32.074 INFO:teuthology.orchestra.run.vm08.stdout:(149/152): protobuf-compiler-3.14.0-17.el9_7.x8 16 MB/s | 862 kB 00:00 2026-04-01T13:53:32.076 INFO:teuthology.orchestra.run.vm06.stdout:(126/152): python3-jinja2-2.11.3-8.el9_5.noarch 7.7 MB/s | 228 kB 00:00 2026-04-01T13:53:32.110 INFO:teuthology.orchestra.run.vm06.stdout:(127/152): python3-jmespath-1.0.1-1.el9_7.noarc 1.3 MB/s | 43 kB 00:00 2026-04-01T13:53:32.225 INFO:teuthology.orchestra.run.vm06.stdout:(128/152): python3-babel-2.9.1-2.el9.noarch.rpm 26 MB/s | 5.8 MB 00:00 2026-04-01T13:53:32.243 INFO:teuthology.orchestra.run.vm06.stdout:(129/152): python3-libstoragemgmt-1.10.1-1.el9. 1.2 MB/s | 166 kB 00:00 2026-04-01T13:53:32.280 INFO:teuthology.orchestra.run.vm06.stdout:(130/152): python3-lxml-4.6.5-3.el9.x86_64.rpm 22 MB/s | 1.2 MB 00:00 2026-04-01T13:53:32.281 INFO:teuthology.orchestra.run.vm06.stdout:(131/152): python3-markupsafe-1.1.1-12.el9.x86_ 841 kB/s | 32 kB 00:00 2026-04-01T13:53:32.315 INFO:teuthology.orchestra.run.vm06.stdout:(132/152): python3-numpy-f2py-1.23.5-2.el9_7.x8 11 MB/s | 368 kB 00:00 2026-04-01T13:53:32.345 INFO:teuthology.orchestra.run.vm06.stdout:(133/152): python3-packaging-20.9-5.el9.noarch. 2.3 MB/s | 69 kB 00:00 2026-04-01T13:53:32.377 INFO:teuthology.orchestra.run.vm06.stdout:(134/152): python3-protobuf-3.14.0-17.el9_7.noa 7.3 MB/s | 237 kB 00:00 2026-04-01T13:53:32.416 INFO:teuthology.orchestra.run.vm06.stdout:(135/152): python3-numpy-1.23.5-2.el9_7.x86_64. 43 MB/s | 5.8 MB 00:00 2026-04-01T13:53:32.417 INFO:teuthology.orchestra.run.vm06.stdout:(136/152): python3-pyasn1-0.4.8-7.el9_7.noarch. 3.2 MB/s | 132 kB 00:00 2026-04-01T13:53:32.435 INFO:teuthology.orchestra.run.vm08.stdout:(150/152): librados2-20.2.0-9.ge2ce8426bc5.el9. 9.0 MB/s | 3.5 MB 00:00 2026-04-01T13:53:32.445 INFO:teuthology.orchestra.run.vm06.stdout:(137/152): python3-pyasn1-modules-0.4.8-7.el9_7 7.1 MB/s | 210 kB 00:00 2026-04-01T13:53:32.487 INFO:teuthology.orchestra.run.vm06.stdout:(138/152): python3-requests-oauthlib-1.3.0-12.e 622 kB/s | 43 kB 00:00 2026-04-01T13:53:32.515 INFO:teuthology.orchestra.run.vm06.stdout:(139/152): python3-toml-0.10.2-6.el9.0.1.noarch 1.5 MB/s | 44 kB 00:00 2026-04-01T13:53:32.545 INFO:teuthology.orchestra.run.vm06.stdout:(140/152): qatlib-24.09.0-1.el9.x86_64.rpm 7.4 MB/s | 221 kB 00:00 2026-04-01T13:53:32.572 INFO:teuthology.orchestra.run.vm06.stdout:(141/152): qatlib-service-24.09.0-1.el9.x86_64. 1.3 MB/s | 36 kB 00:00 2026-04-01T13:53:32.603 INFO:teuthology.orchestra.run.vm06.stdout:(142/152): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.1 MB/s | 65 kB 00:00 2026-04-01T13:53:32.635 INFO:teuthology.orchestra.run.vm06.stdout:(143/152): socat-1.7.4.1-8.el9.x86_64.rpm 9.1 MB/s | 299 kB 00:00 2026-04-01T13:53:32.666 INFO:teuthology.orchestra.run.vm06.stdout:(144/152): xmlsec1-1.2.29-13.el9.x86_64.rpm 6.0 MB/s | 188 kB 00:00 2026-04-01T13:53:32.739 INFO:teuthology.orchestra.run.vm06.stdout:(145/152): python3-scipy-1.9.3-2.el9.x86_64.rpm 64 MB/s | 19 MB 00:00 2026-04-01T13:53:32.740 INFO:teuthology.orchestra.run.vm08.stdout:(151/152): librbd1-20.2.0-9.ge2ce8426bc5.el9.cl 4.3 MB/s | 2.8 MB 00:00 2026-04-01T13:53:32.741 INFO:teuthology.orchestra.run.vm06.stdout:(146/152): xmlsec1-openssl-1.2.29-13.el9.x86_64 1.2 MB/s | 89 kB 00:00 2026-04-01T13:53:32.767 INFO:teuthology.orchestra.run.vm06.stdout:(147/152): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.3 MB/s | 63 kB 00:00 2026-04-01T13:53:32.799 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-04-01T13:53:32.801 INFO:teuthology.orchestra.run.vm06.stdout:(148/152): lua-devel-5.4.4-4.el9.x86_64.rpm 357 kB/s | 21 kB 00:00 2026-04-01T13:53:32.806 INFO:teuthology.orchestra.run.vm06.stdout:(149/152): protobuf-compiler-3.14.0-17.el9_7.x8 21 MB/s | 862 kB 00:00 2026-04-01T13:53:32.854 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-04-01T13:53:32.854 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-04-01T13:53:33.212 INFO:teuthology.orchestra.run.vm06.stdout:(150/152): librados2-20.2.0-9.ge2ce8426bc5.el9. 8.5 MB/s | 3.5 MB 00:00 2026-04-01T13:53:33.254 INFO:teuthology.orchestra.run.vm06.stdout:(151/152): librbd1-20.2.0-9.ge2ce8426bc5.el9.cl 6.4 MB/s | 2.8 MB 00:00 2026-04-01T13:53:33.885 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-04-01T13:53:33.885 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-04-01T13:53:34.550 INFO:teuthology.orchestra.run.vm08.stdout:(152/152): python3-scipy-1.9.3-2.el9.x86_64.rpm 6.2 MB/s | 19 MB 00:03 2026-04-01T13:53:34.555 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-04-01T13:53:34.555 INFO:teuthology.orchestra.run.vm08.stdout:Total 21 MB/s | 274 MB 00:12 2026-04-01T13:53:34.768 INFO:teuthology.orchestra.run.vm06.stdout:(152/152): ceph-test-20.2.0-9.ge2ce8426bc5.el9. 9.5 MB/s | 85 MB 00:08 2026-04-01T13:53:34.768 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-04-01T13:53:34.768 INFO:teuthology.orchestra.run.vm06.stdout:Total 23 MB/s | 274 MB 00:11 2026-04-01T13:53:35.021 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-04-01T13:53:35.031 INFO:teuthology.orchestra.run.vm09.stdout: Installing : thrift-0.15.0-4.el9.x86_64 1/154 2026-04-01T13:53:35.045 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 2/154 2026-04-01T13:53:35.228 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/154 2026-04-01T13:53:35.231 INFO:teuthology.orchestra.run.vm09.stdout: Upgrading : librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:53:35.269 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-04-01T13:53:35.291 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:53:35.294 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 5/154 2026-04-01T13:53:35.310 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 5/154 2026-04-01T13:53:35.314 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/154 2026-04-01T13:53:35.315 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 7/154 2026-04-01T13:53:35.330 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-04-01T13:53:35.330 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-04-01T13:53:35.350 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 7/154 2026-04-01T13:53:35.357 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 8/154 2026-04-01T13:53:35.368 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libxslt-1.1.34-13.el9_6.x86_64 9/154 2026-04-01T13:53:35.372 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 10/154 2026-04-01T13:53:35.377 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 11/154 2026-04-01T13:53:35.383 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 12/154 2026-04-01T13:53:35.387 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lmdb-libs-0.9.29-3.el9.x86_64 13/154 2026-04-01T13:53:35.393 INFO:teuthology.orchestra.run.vm09.stdout: Installing : liboath-2.6.12-1.el9.x86_64 14/154 2026-04-01T13:53:35.540 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 15/154 2026-04-01T13:53:35.543 INFO:teuthology.orchestra.run.vm09.stdout: Upgrading : librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 16/154 2026-04-01T13:53:35.562 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 16/154 2026-04-01T13:53:35.584 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-04-01T13:53:35.604 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 17/154 2026-04-01T13:53:35.613 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 18/154 2026-04-01T13:53:35.623 INFO:teuthology.orchestra.run.vm09.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 19/154 2026-04-01T13:53:35.624 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 20/154 2026-04-01T13:53:35.644 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-04-01T13:53:35.644 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-04-01T13:53:35.652 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 20/154 2026-04-01T13:53:35.653 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 21/154 2026-04-01T13:53:35.699 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 21/154 2026-04-01T13:53:35.713 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyasn1-0.4.8-7.el9_7.noarch 22/154 2026-04-01T13:53:35.721 INFO:teuthology.orchestra.run.vm09.stdout: Installing : protobuf-3.14.0-17.el9_7.x86_64 23/154 2026-04-01T13:53:35.725 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/154 2026-04-01T13:53:35.732 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-3.0.4-8.el9.0.1.x86_64 25/154 2026-04-01T13:53:35.761 INFO:teuthology.orchestra.run.vm09.stdout: Installing : unzip-6.0-59.el9.x86_64 26/154 2026-04-01T13:53:35.777 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urllib3-1.26.5-6.el9_7.1.noarch 27/154 2026-04-01T13:53:35.782 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-requests-2.25.1-10.el9_6.noarch 28/154 2026-04-01T13:53:35.790 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libquadmath-11.5.0-11.el9.x86_64 29/154 2026-04-01T13:53:35.793 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libgfortran-11.5.0-11.el9.x86_64 30/154 2026-04-01T13:53:35.797 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/154 2026-04-01T13:53:35.833 INFO:teuthology.orchestra.run.vm09.stdout: Installing : re2-1:20211101-20.el9.x86_64 32/154 2026-04-01T13:53:35.868 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 33/154 2026-04-01T13:53:35.875 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 34/154 2026-04-01T13:53:35.886 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el 35/154 2026-04-01T13:53:35.900 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso 36/154 2026-04-01T13:53:35.908 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 37/154 2026-04-01T13:53:35.938 INFO:teuthology.orchestra.run.vm09.stdout: Installing : zip-3.0-35.el9.x86_64 38/154 2026-04-01T13:53:35.944 INFO:teuthology.orchestra.run.vm09.stdout: Installing : luarocks-3.9.2-5.el9.noarch 39/154 2026-04-01T13:53:35.953 INFO:teuthology.orchestra.run.vm09.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 40/154 2026-04-01T13:53:35.970 INFO:teuthology.orchestra.run.vm09.stdout: Installing : protobuf-compiler-3.14.0-17.el9_7.x86_64 41/154 2026-04-01T13:53:36.029 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/154 2026-04-01T13:53:36.035 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 43/154 2026-04-01T13:53:36.042 INFO:teuthology.orchestra.run.vm09.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 44/154 2026-04-01T13:53:36.049 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 45/154 2026-04-01T13:53:36.068 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 46/154 2026-04-01T13:53:36.073 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 47/154 2026-04-01T13:53:36.082 INFO:teuthology.orchestra.run.vm09.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 48/154 2026-04-01T13:53:36.093 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso 49/154 2026-04-01T13:53:36.101 INFO:teuthology.orchestra.run.vm09.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 50/154 2026-04-01T13:53:36.107 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-toml-0.10.2-6.el9.0.1.noarch 51/154 2026-04-01T13:53:36.115 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 52/154 2026-04-01T13:53:36.121 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 53/154 2026-04-01T13:53:36.155 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 54/154 2026-04-01T13:53:36.165 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-protobuf-3.14.0-17.el9_7.noarch 55/154 2026-04-01T13:53:36.174 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 56/154 2026-04-01T13:53:36.217 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jmespath-1.0.1-1.el9_7.noarch 57/154 2026-04-01T13:53:36.376 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-04-01T13:53:36.376 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-04-01T13:53:36.502 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-devel-3.9.23-2.el9.x86_64 58/154 2026-04-01T13:53:36.535 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 59/154 2026-04-01T13:53:36.540 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jinja2-2.11.3-8.el9_5.noarch 60/154 2026-04-01T13:53:36.544 INFO:teuthology.orchestra.run.vm09.stdout: Installing : perl-Benchmark-1.23-481.1.el9_6.noarch 61/154 2026-04-01T13:53:36.614 INFO:teuthology.orchestra.run.vm09.stdout: Installing : openblas-0.3.29-1.el9.x86_64 62/154 2026-04-01T13:53:36.617 INFO:teuthology.orchestra.run.vm09.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 63/154 2026-04-01T13:53:36.642 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 64/154 2026-04-01T13:53:36.754 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-04-01T13:53:36.754 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-04-01T13:53:37.061 INFO:teuthology.orchestra.run.vm09.stdout: Installing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 65/154 2026-04-01T13:53:37.163 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-numpy-1:1.23.5-2.el9_7.x86_64 66/154 2026-04-01T13:53:37.495 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-04-01T13:53:37.505 INFO:teuthology.orchestra.run.vm08.stdout: Installing : thrift-0.15.0-4.el9.x86_64 1/154 2026-04-01T13:53:37.520 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 2/154 2026-04-01T13:53:37.710 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/154 2026-04-01T13:53:37.712 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:53:37.774 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:53:37.844 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 5/154 2026-04-01T13:53:37.909 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 5/154 2026-04-01T13:53:37.936 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-04-01T13:53:37.961 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/154 2026-04-01T13:53:37.994 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 7/154 2026-04-01T13:53:38.028 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 67/154 2026-04-01T13:53:38.047 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 1/154 2026-04-01T13:53:38.095 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 68/154 2026-04-01T13:53:38.097 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 2/154 2026-04-01T13:53:38.121 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 7/154 2026-04-01T13:53:38.135 INFO:teuthology.orchestra.run.vm09.stdout: Installing : boost-program-options-1.75.0-13.el9_7.x86_64 69/154 2026-04-01T13:53:38.156 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 8/154 2026-04-01T13:53:38.168 INFO:teuthology.orchestra.run.vm09.stdout: Installing : smartmontools-1:7.2-9.el9.x86_64 70/154 2026-04-01T13:53:38.185 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: smartmontools-1:7.2-9.el9.x86_64 70/154 2026-04-01T13:53:38.185 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-04-01T13:53:38.185 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:38.202 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libxslt-1.1.34-13.el9_6.x86_64 9/154 2026-04-01T13:53:38.214 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 71/154 2026-04-01T13:53:38.262 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 10/154 2026-04-01T13:53:38.267 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-packaging-20.9-5.el9.noarch 72/154 2026-04-01T13:53:38.269 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 11/154 2026-04-01T13:53:38.274 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 12/154 2026-04-01T13:53:38.277 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lmdb-libs-0.9.29-3.el9.x86_64 13/154 2026-04-01T13:53:38.283 INFO:teuthology.orchestra.run.vm08.stdout: Installing : liboath-2.6.12-1.el9.x86_64 14/154 2026-04-01T13:53:38.288 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-ply-3.11-14.el9.0.1.noarch 73/154 2026-04-01T13:53:38.316 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 74/154 2026-04-01T13:53:38.345 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/154 2026-04-01T13:53:38.347 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:53:38.415 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 75/154 2026-04-01T13:53:38.438 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 15/154 2026-04-01T13:53:38.483 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 16/154 2026-04-01T13:53:38.492 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cryptography-36.0.1-5.el9_6.x86_64 76/154 2026-04-01T13:53:38.523 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 16/154 2026-04-01T13:53:38.549 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 77/154 2026-04-01T13:53:38.562 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:53:38.564 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cheroot-10.0.1-5.el9.noarch 78/154 2026-04-01T13:53:38.564 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 5/154 2026-04-01T13:53:38.568 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 17/154 2026-04-01T13:53:38.570 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 79/154 2026-04-01T13:53:38.573 INFO:teuthology.orchestra.run.vm09.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 80/154 2026-04-01T13:53:38.577 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 18/154 2026-04-01T13:53:38.582 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 5/154 2026-04-01T13:53:38.585 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/154 2026-04-01T13:53:38.587 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 19/154 2026-04-01T13:53:38.588 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 7/154 2026-04-01T13:53:38.589 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 20/154 2026-04-01T13:53:38.604 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-24.09.0-1.el9.x86_64 81/154 2026-04-01T13:53:38.609 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatlib-24.09.0-1.el9.x86_64 81/154 2026-04-01T13:53:38.611 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatlib-service-24.09.0-1.el9.x86_64 82/154 2026-04-01T13:53:38.615 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 20/154 2026-04-01T13:53:38.617 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 21/154 2026-04-01T13:53:38.622 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 7/154 2026-04-01T13:53:38.630 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 8/154 2026-04-01T13:53:38.632 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: qatlib-service-24.09.0-1.el9.x86_64 82/154 2026-04-01T13:53:38.641 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-13.el9_6.x86_64 9/154 2026-04-01T13:53:38.646 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 10/154 2026-04-01T13:53:38.651 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 11/154 2026-04-01T13:53:38.655 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 12/154 2026-04-01T13:53:38.657 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 21/154 2026-04-01T13:53:38.658 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lmdb-libs-0.9.29-3.el9.x86_64 13/154 2026-04-01T13:53:38.664 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 14/154 2026-04-01T13:53:38.670 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-0.4.8-7.el9_7.noarch 22/154 2026-04-01T13:53:38.677 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-3.14.0-17.el9_7.x86_64 23/154 2026-04-01T13:53:38.681 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/154 2026-04-01T13:53:38.687 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-3.0.4-8.el9.0.1.x86_64 25/154 2026-04-01T13:53:38.719 INFO:teuthology.orchestra.run.vm08.stdout: Installing : unzip-6.0-59.el9.x86_64 26/154 2026-04-01T13:53:38.735 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urllib3-1.26.5-6.el9_7.1.noarch 27/154 2026-04-01T13:53:38.740 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-2.25.1-10.el9_6.noarch 28/154 2026-04-01T13:53:38.748 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libquadmath-11.5.0-11.el9.x86_64 29/154 2026-04-01T13:53:38.750 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libgfortran-11.5.0-11.el9.x86_64 30/154 2026-04-01T13:53:38.774 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/154 2026-04-01T13:53:38.781 INFO:teuthology.orchestra.run.vm09.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 83/154 2026-04-01T13:53:38.785 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvme-cli-2.13-1.el9.x86_64 84/154 2026-04-01T13:53:38.813 INFO:teuthology.orchestra.run.vm08.stdout: Installing : re2-1:20211101-20.el9.x86_64 32/154 2026-04-01T13:53:38.815 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 15/154 2026-04-01T13:53:38.818 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 16/154 2026-04-01T13:53:38.835 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 16/154 2026-04-01T13:53:38.853 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 33/154 2026-04-01T13:53:38.861 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 34/154 2026-04-01T13:53:38.874 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el 35/154 2026-04-01T13:53:38.893 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 17/154 2026-04-01T13:53:38.893 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso 36/154 2026-04-01T13:53:38.902 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 18/154 2026-04-01T13:53:38.905 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 37/154 2026-04-01T13:53:38.914 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 19/154 2026-04-01T13:53:38.916 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 20/154 2026-04-01T13:53:38.935 INFO:teuthology.orchestra.run.vm08.stdout: Installing : zip-3.0-35.el9.x86_64 38/154 2026-04-01T13:53:38.941 INFO:teuthology.orchestra.run.vm08.stdout: Installing : luarocks-3.9.2-5.el9.noarch 39/154 2026-04-01T13:53:38.946 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 20/154 2026-04-01T13:53:38.948 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 21/154 2026-04-01T13:53:38.952 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 40/154 2026-04-01T13:53:38.968 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-compiler-3.14.0-17.el9_7.x86_64 41/154 2026-04-01T13:53:38.992 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 21/154 2026-04-01T13:53:39.008 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9_7.noarch 22/154 2026-04-01T13:53:39.018 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9_7.x86_64 23/154 2026-04-01T13:53:39.023 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 24/154 2026-04-01T13:53:39.031 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-8.el9.0.1.x86_64 25/154 2026-04-01T13:53:39.034 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/154 2026-04-01T13:53:39.040 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 43/154 2026-04-01T13:53:39.047 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 44/154 2026-04-01T13:53:39.054 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 45/154 2026-04-01T13:53:39.064 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 26/154 2026-04-01T13:53:39.074 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 46/154 2026-04-01T13:53:39.079 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 47/154 2026-04-01T13:53:39.083 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-6.el9_7.1.noarch 27/154 2026-04-01T13:53:39.088 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9_6.noarch 28/154 2026-04-01T13:53:39.090 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 48/154 2026-04-01T13:53:39.096 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-11.el9.x86_64 29/154 2026-04-01T13:53:39.100 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvme-cli-2.13-1.el9.x86_64 84/154 2026-04-01T13:53:39.100 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-04-01T13:53:39.100 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:39.100 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-11.el9.x86_64 30/154 2026-04-01T13:53:39.102 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso 49/154 2026-04-01T13:53:39.105 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 31/154 2026-04-01T13:53:39.111 INFO:teuthology.orchestra.run.vm08.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 50/154 2026-04-01T13:53:39.112 INFO:teuthology.orchestra.run.vm09.stdout: Installing : mailcap-2.1.49-5.el9.0.2.noarch 85/154 2026-04-01T13:53:39.117 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-toml-0.10.2-6.el9.0.1.noarch 51/154 2026-04-01T13:53:39.117 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 86/154 2026-04-01T13:53:39.126 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 52/154 2026-04-01T13:53:39.132 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 53/154 2026-04-01T13:53:39.139 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:39.139 INFO:teuthology.orchestra.run.vm09.stdout:Creating group 'libstoragemgmt' with GID 992. 2026-04-01T13:53:39.139 INFO:teuthology.orchestra.run.vm09.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 992 and GID 992. 2026-04-01T13:53:39.139 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:39.145 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 32/154 2026-04-01T13:53:39.150 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:39.169 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 54/154 2026-04-01T13:53:39.180 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-protobuf-3.14.0-17.el9_7.noarch 55/154 2026-04-01T13:53:39.184 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:39.184 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-04-01T13:53:39.184 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:39.189 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 56/154 2026-04-01T13:53:39.192 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 33/154 2026-04-01T13:53:39.201 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 34/154 2026-04-01T13:53:39.233 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jmespath-1.0.1-1.el9_7.noarch 57/154 2026-04-01T13:53:39.241 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 88/154 2026-04-01T13:53:39.246 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el 35/154 2026-04-01T13:53:39.261 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso 36/154 2026-04-01T13:53:39.270 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 37/154 2026-04-01T13:53:39.271 INFO:teuthology.orchestra.run.vm09.stdout: Installing : fuse-2.9.9-17.el9.x86_64 89/154 2026-04-01T13:53:39.277 INFO:teuthology.orchestra.run.vm09.stdout: Installing : cryptsetup-2.7.2-4.el9.x86_64 90/154 2026-04-01T13:53:39.285 INFO:teuthology.orchestra.run.vm09.stdout: Installing : c-ares-1.19.1-2.el9_4.x86_64 91/154 2026-04-01T13:53:39.290 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 92/154 2026-04-01T13:53:39.303 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 38/154 2026-04-01T13:53:39.310 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 39/154 2026-04-01T13:53:39.311 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 93/154 2026-04-01T13:53:39.319 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 94/154 2026-04-01T13:53:39.320 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 40/154 2026-04-01T13:53:39.337 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9_7.x86_64 41/154 2026-04-01T13:53:39.388 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 95/154 2026-04-01T13:53:39.401 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 96/154 2026-04-01T13:53:39.408 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 42/154 2026-04-01T13:53:39.413 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 43/154 2026-04-01T13:53:39.417 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 97/154 2026-04-01T13:53:39.421 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 44/154 2026-04-01T13:53:39.428 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 45/154 2026-04-01T13:53:39.430 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 98/154 2026-04-01T13:53:39.440 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 99/154 2026-04-01T13:53:39.449 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 46/154 2026-04-01T13:53:39.455 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 47/154 2026-04-01T13:53:39.465 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 100/154 2026-04-01T13:53:39.466 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 48/154 2026-04-01T13:53:39.477 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso 49/154 2026-04-01T13:53:39.486 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 50/154 2026-04-01T13:53:39.488 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 101/154 2026-04-01T13:53:39.493 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.0.1.noarch 51/154 2026-04-01T13:53:39.503 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 52/154 2026-04-01T13:53:39.509 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 53/154 2026-04-01T13:53:39.511 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 102/154 2026-04-01T13:53:39.518 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 103/154 2026-04-01T13:53:39.527 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-devel-3.9.23-2.el9.x86_64 58/154 2026-04-01T13:53:39.529 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 104/154 2026-04-01T13:53:39.545 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 54/154 2026-04-01T13:53:39.554 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9_7.noarch 55/154 2026-04-01T13:53:39.560 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 59/154 2026-04-01T13:53:39.565 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jinja2-2.11.3-8.el9_5.noarch 60/154 2026-04-01T13:53:39.565 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 56/154 2026-04-01T13:53:39.569 INFO:teuthology.orchestra.run.vm08.stdout: Installing : perl-Benchmark-1.23-481.1.el9_6.noarch 61/154 2026-04-01T13:53:39.582 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 105/154 2026-04-01T13:53:39.614 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9_7.noarch 57/154 2026-04-01T13:53:39.635 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-0.3.29-1.el9.x86_64 62/154 2026-04-01T13:53:39.638 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 63/154 2026-04-01T13:53:39.666 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 64/154 2026-04-01T13:53:39.925 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.23-2.el9.x86_64 58/154 2026-04-01T13:53:39.960 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 59/154 2026-04-01T13:53:39.965 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9_5.noarch 60/154 2026-04-01T13:53:39.971 INFO:teuthology.orchestra.run.vm06.stdout: Installing : perl-Benchmark-1.23-481.1.el9_6.noarch 61/154 2026-04-01T13:53:39.993 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 106/154 2026-04-01T13:53:40.012 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 107/154 2026-04-01T13:53:40.018 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 108/154 2026-04-01T13:53:40.027 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 109/154 2026-04-01T13:53:40.036 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 110/154 2026-04-01T13:53:40.042 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 62/154 2026-04-01T13:53:40.042 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 111/154 2026-04-01T13:53:40.045 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 63/154 2026-04-01T13:53:40.064 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 65/154 2026-04-01T13:53:40.073 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 64/154 2026-04-01T13:53:40.079 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 112/154 2026-04-01T13:53:40.086 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-cherrypy-18.10.0-5.el9.noarch 113/154 2026-04-01T13:53:40.094 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 114/154 2026-04-01T13:53:40.099 INFO:teuthology.orchestra.run.vm09.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 115/154 2026-04-01T13:53:40.107 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 116/154 2026-04-01T13:53:40.158 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-1:1.23.5-2.el9_7.x86_64 66/154 2026-04-01T13:53:40.440 INFO:teuthology.orchestra.run.vm09.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 117/154 2026-04-01T13:53:40.442 INFO:teuthology.orchestra.run.vm09.stdout: Installing : librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 118/154 2026-04-01T13:53:40.464 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 118/154 2026-04-01T13:53:40.466 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 119/154 2026-04-01T13:53:40.492 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 65/154 2026-04-01T13:53:40.581 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9_7.x86_64 66/154 2026-04-01T13:53:40.994 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 67/154 2026-04-01T13:53:41.018 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 68/154 2026-04-01T13:53:41.034 INFO:teuthology.orchestra.run.vm08.stdout: Installing : boost-program-options-1.75.0-13.el9_7.x86_64 69/154 2026-04-01T13:53:41.036 INFO:teuthology.orchestra.run.vm08.stdout: Installing : smartmontools-1:7.2-9.el9.x86_64 70/154 2026-04-01T13:53:41.051 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: smartmontools-1:7.2-9.el9.x86_64 70/154 2026-04-01T13:53:41.051 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-04-01T13:53:41.051 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:41.074 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 71/154 2026-04-01T13:53:41.083 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-packaging-20.9-5.el9.noarch 72/154 2026-04-01T13:53:41.101 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ply-3.11-14.el9.0.1.noarch 73/154 2026-04-01T13:53:41.123 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 74/154 2026-04-01T13:53:41.215 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 75/154 2026-04-01T13:53:41.229 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cryptography-36.0.1-5.el9_6.x86_64 76/154 2026-04-01T13:53:41.259 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 77/154 2026-04-01T13:53:41.268 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cheroot-10.0.1-5.el9.noarch 78/154 2026-04-01T13:53:41.276 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 79/154 2026-04-01T13:53:41.279 INFO:teuthology.orchestra.run.vm08.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 80/154 2026-04-01T13:53:41.309 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-24.09.0-1.el9.x86_64 81/154 2026-04-01T13:53:41.314 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-24.09.0-1.el9.x86_64 81/154 2026-04-01T13:53:41.316 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-service-24.09.0-1.el9.x86_64 82/154 2026-04-01T13:53:41.336 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-24.09.0-1.el9.x86_64 82/154 2026-04-01T13:53:41.418 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 67/154 2026-04-01T13:53:41.440 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 68/154 2026-04-01T13:53:41.454 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9_7.x86_64 69/154 2026-04-01T13:53:41.457 INFO:teuthology.orchestra.run.vm06.stdout: Installing : smartmontools-1:7.2-9.el9.x86_64 70/154 2026-04-01T13:53:41.472 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 83/154 2026-04-01T13:53:41.474 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: smartmontools-1:7.2-9.el9.x86_64 70/154 2026-04-01T13:53:41.475 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-04-01T13:53:41.475 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:41.476 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvme-cli-2.13-1.el9.x86_64 84/154 2026-04-01T13:53:41.495 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 71/154 2026-04-01T13:53:41.504 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 72/154 2026-04-01T13:53:41.521 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.0.1.noarch 73/154 2026-04-01T13:53:41.541 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 74/154 2026-04-01T13:53:41.642 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 75/154 2026-04-01T13:53:41.656 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9_6.x86_64 76/154 2026-04-01T13:53:41.686 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 77/154 2026-04-01T13:53:41.694 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-5.el9.noarch 78/154 2026-04-01T13:53:41.700 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 79/154 2026-04-01T13:53:41.703 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 80/154 2026-04-01T13:53:41.736 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-24.09.0-1.el9.x86_64 81/154 2026-04-01T13:53:41.740 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-24.09.0-1.el9.x86_64 81/154 2026-04-01T13:53:41.742 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-24.09.0-1.el9.x86_64 82/154 2026-04-01T13:53:41.765 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-24.09.0-1.el9.x86_64 82/154 2026-04-01T13:53:41.765 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:41.771 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:41.794 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvme-cli-2.13-1.el9.x86_64 84/154 2026-04-01T13:53:41.794 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-04-01T13:53:41.794 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:41.804 INFO:teuthology.orchestra.run.vm08.stdout: Installing : mailcap-2.1.49-5.el9.0.2.noarch 85/154 2026-04-01T13:53:41.807 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 86/154 2026-04-01T13:53:41.832 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:41.833 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'libstoragemgmt' with GID 992. 2026-04-01T13:53:41.833 INFO:teuthology.orchestra.run.vm08.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 992 and GID 992. 2026-04-01T13:53:41.833 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:41.852 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:41.881 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:41.881 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-04-01T13:53:41.881 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:41.899 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 83/154 2026-04-01T13:53:41.904 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 88/154 2026-04-01T13:53:41.904 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.13-1.el9.x86_64 84/154 2026-04-01T13:53:41.931 INFO:teuthology.orchestra.run.vm08.stdout: Installing : fuse-2.9.9-17.el9.x86_64 89/154 2026-04-01T13:53:41.939 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cryptsetup-2.7.2-4.el9.x86_64 90/154 2026-04-01T13:53:41.946 INFO:teuthology.orchestra.run.vm08.stdout: Installing : c-ares-1.19.1-2.el9_4.x86_64 91/154 2026-04-01T13:53:41.951 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 92/154 2026-04-01T13:53:41.969 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 93/154 2026-04-01T13:53:41.977 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 94/154 2026-04-01T13:53:42.039 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 95/154 2026-04-01T13:53:42.052 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 96/154 2026-04-01T13:53:42.068 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 97/154 2026-04-01T13:53:42.080 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 98/154 2026-04-01T13:53:42.090 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 99/154 2026-04-01T13:53:42.116 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 100/154 2026-04-01T13:53:42.128 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:42.135 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 121/154 2026-04-01T13:53:42.136 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 101/154 2026-04-01T13:53:42.156 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 102/154 2026-04-01T13:53:42.163 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 103/154 2026-04-01T13:53:42.173 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 104/154 2026-04-01T13:53:42.189 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 121/154 2026-04-01T13:53:42.189 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-04-01T13:53:42.189 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-04-01T13:53:42.189 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:42.194 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 122/154 2026-04-01T13:53:42.224 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 105/154 2026-04-01T13:53:42.225 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.13-1.el9.x86_64 84/154 2026-04-01T13:53:42.225 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-04-01T13:53:42.225 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:42.233 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.0.2.noarch 85/154 2026-04-01T13:53:42.235 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 86/154 2026-04-01T13:53:42.265 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:42.265 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 992. 2026-04-01T13:53:42.265 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 992 and GID 992. 2026-04-01T13:53:42.265 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:42.275 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:42.315 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 87/154 2026-04-01T13:53:42.315 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-04-01T13:53:42.315 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:42.337 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 88/154 2026-04-01T13:53:42.365 INFO:teuthology.orchestra.run.vm06.stdout: Installing : fuse-2.9.9-17.el9.x86_64 89/154 2026-04-01T13:53:42.371 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.7.2-4.el9.x86_64 90/154 2026-04-01T13:53:42.381 INFO:teuthology.orchestra.run.vm06.stdout: Installing : c-ares-1.19.1-2.el9_4.x86_64 91/154 2026-04-01T13:53:42.388 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 92/154 2026-04-01T13:53:42.409 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 93/154 2026-04-01T13:53:42.418 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 94/154 2026-04-01T13:53:42.492 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 95/154 2026-04-01T13:53:42.504 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 96/154 2026-04-01T13:53:42.521 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 97/154 2026-04-01T13:53:42.537 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 98/154 2026-04-01T13:53:42.548 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 99/154 2026-04-01T13:53:42.575 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 100/154 2026-04-01T13:53:42.599 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 101/154 2026-04-01T13:53:42.621 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 102/154 2026-04-01T13:53:42.625 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 106/154 2026-04-01T13:53:42.629 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 103/154 2026-04-01T13:53:42.641 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 104/154 2026-04-01T13:53:42.643 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 107/154 2026-04-01T13:53:42.650 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 108/154 2026-04-01T13:53:42.658 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 109/154 2026-04-01T13:53:42.667 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 110/154 2026-04-01T13:53:42.674 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 111/154 2026-04-01T13:53:42.695 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 105/154 2026-04-01T13:53:42.711 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 112/154 2026-04-01T13:53:42.721 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cherrypy-18.10.0-5.el9.noarch 113/154 2026-04-01T13:53:42.730 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 114/154 2026-04-01T13:53:42.735 INFO:teuthology.orchestra.run.vm08.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 115/154 2026-04-01T13:53:42.743 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 116/154 2026-04-01T13:53:43.073 INFO:teuthology.orchestra.run.vm08.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 117/154 2026-04-01T13:53:43.077 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 118/154 2026-04-01T13:53:43.097 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 118/154 2026-04-01T13:53:43.100 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 119/154 2026-04-01T13:53:43.120 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 106/154 2026-04-01T13:53:43.137 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 107/154 2026-04-01T13:53:43.143 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 108/154 2026-04-01T13:53:43.152 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 109/154 2026-04-01T13:53:43.160 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 110/154 2026-04-01T13:53:43.167 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 111/154 2026-04-01T13:53:43.203 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 112/154 2026-04-01T13:53:43.211 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.10.0-5.el9.noarch 113/154 2026-04-01T13:53:43.219 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 114/154 2026-04-01T13:53:43.224 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 115/154 2026-04-01T13:53:43.232 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 116/154 2026-04-01T13:53:43.580 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 117/154 2026-04-01T13:53:43.584 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 118/154 2026-04-01T13:53:43.609 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 118/154 2026-04-01T13:53:43.611 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 119/154 2026-04-01T13:53:44.394 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:44.619 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:44.938 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:44.964 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:45.061 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 121/154 2026-04-01T13:53:45.094 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:45.168 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 121/154 2026-04-01T13:53:45.168 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-04-01T13:53:45.168 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-04-01T13:53:45.168 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:45.174 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 122/154 2026-04-01T13:53:45.460 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 120/154 2026-04-01T13:53:45.468 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 121/154 2026-04-01T13:53:45.513 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 121/154 2026-04-01T13:53:45.513 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-04-01T13:53:45.513 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-04-01T13:53:45.513 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:45.519 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 122/154 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 122/154 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /sys 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /proc 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /mnt 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /var/tmp 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /home 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /root 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout:skipping the directory /tmp 2026-04-01T13:53:48.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:49.090 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 123/154 2026-04-01T13:53:49.108 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 123/154 2026-04-01T13:53:49.108 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:49.108 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-01T13:53:49.108 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:49.917 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 124/154 2026-04-01T13:53:49.946 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 124/154 2026-04-01T13:53:49.946 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:49.946 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-01T13:53:49.946 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-01T13:53:49.946 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-01T13:53:49.946 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:50.056 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 125/154 2026-04-01T13:53:50.081 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 125/154 2026-04-01T13:53:50.081 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:50.081 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-01T13:53:50.081 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-01T13:53:50.081 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-01T13:53:50.081 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:50.385 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 126/154 2026-04-01T13:53:50.411 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 126/154 2026-04-01T13:53:50.411 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:50.411 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-01T13:53:50.411 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-01T13:53:50.411 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-01T13:53:50.411 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:50.451 INFO:teuthology.orchestra.run.vm09.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 127/154 2026-04-01T13:53:50.519 INFO:teuthology.orchestra.run.vm09.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 128/154 2026-04-01T13:53:50.658 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 129/154 2026-04-01T13:53:50.714 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 130/154 2026-04-01T13:53:51.645 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 131/154 2026-04-01T13:53:51.769 INFO:teuthology.orchestra.run.vm09.stdout: Installing : cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 131/154 2026-04-01T13:53:52.143 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 132/154 2026-04-01T13:53:52.239 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 133/154 2026-04-01T13:53:52.326 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 134/154 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 122/154 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-04-01T13:53:52.802 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:52.915 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 123/154 2026-04-01T13:53:52.931 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 123/154 2026-04-01T13:53:52.931 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:52.931 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-01T13:53:52.931 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:53.207 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 122/154 2026-04-01T13:53:53.207 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-04-01T13:53:53.207 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-04-01T13:53:53.208 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-04-01T13:53:53.208 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-04-01T13:53:53.208 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-04-01T13:53:53.208 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-04-01T13:53:53.208 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-04-01T13:53:53.208 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:53.296 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 123/154 2026-04-01T13:53:53.315 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 123/154 2026-04-01T13:53:53.315 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:53.315 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-01T13:53:53.315 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:53.776 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 124/154 2026-04-01T13:53:53.810 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 124/154 2026-04-01T13:53:53.810 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:53.811 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-01T13:53:53.811 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-01T13:53:53.811 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-01T13:53:53.811 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:53.924 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 125/154 2026-04-01T13:53:53.950 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 125/154 2026-04-01T13:53:53.951 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:53.951 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-01T13:53:53.951 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-01T13:53:53.951 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-01T13:53:53.951 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:54.183 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 124/154 2026-04-01T13:53:54.200 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 134/154 2026-04-01T13:53:54.207 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 126/154 2026-04-01T13:53:54.212 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 124/154 2026-04-01T13:53:54.212 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:54.212 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-01T13:53:54.212 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-01T13:53:54.212 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-01T13:53:54.212 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:54.215 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 135/154 2026-04-01T13:53:54.232 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 126/154 2026-04-01T13:53:54.232 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:54.232 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-01T13:53:54.232 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-01T13:53:54.232 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-01T13:53:54.232 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:54.268 INFO:teuthology.orchestra.run.vm08.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 127/154 2026-04-01T13:53:54.321 INFO:teuthology.orchestra.run.vm08.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 128/154 2026-04-01T13:53:54.335 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 129/154 2026-04-01T13:53:54.340 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 130/154 2026-04-01T13:53:54.340 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 125/154 2026-04-01T13:53:54.368 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 125/154 2026-04-01T13:53:54.368 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:54.368 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-01T13:53:54.368 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-01T13:53:54.368 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-01T13:53:54.368 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:54.409 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 131/154 2026-04-01T13:53:54.414 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 131/154 2026-04-01T13:53:54.422 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 132/154 2026-04-01T13:53:54.454 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 133/154 2026-04-01T13:53:54.458 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 134/154 2026-04-01T13:53:54.650 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 126/154 2026-04-01T13:53:54.676 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 126/154 2026-04-01T13:53:54.676 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:54.676 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-01T13:53:54.676 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-01T13:53:54.676 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-01T13:53:54.676 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:54.714 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 127/154 2026-04-01T13:53:54.775 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 128/154 2026-04-01T13:53:54.782 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 135/154 2026-04-01T13:53:54.785 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 136/154 2026-04-01T13:53:54.790 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 129/154 2026-04-01T13:53:54.793 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 130/154 2026-04-01T13:53:54.800 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 136/154 2026-04-01T13:53:54.801 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 137/154 2026-04-01T13:53:54.859 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 131/154 2026-04-01T13:53:54.863 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 131/154 2026-04-01T13:53:54.867 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 137/154 2026-04-01T13:53:54.872 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 132/154 2026-04-01T13:53:54.903 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 133/154 2026-04-01T13:53:54.907 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 134/154 2026-04-01T13:53:54.919 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 138/154 2026-04-01T13:53:54.921 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 139/154 2026-04-01T13:53:54.946 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 139/154 2026-04-01T13:53:54.946 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:54.946 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-01T13:53:54.946 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-01T13:53:54.946 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-01T13:53:54.946 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:54.962 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 140/154 2026-04-01T13:53:54.974 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 140/154 2026-04-01T13:53:56.184 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 141/154 2026-04-01T13:53:56.188 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 142/154 2026-04-01T13:53:56.219 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 142/154 2026-04-01T13:53:56.219 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:56.219 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-01T13:53:56.219 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-01T13:53:56.219 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-01T13:53:56.219 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:56.233 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 143/154 2026-04-01T13:53:56.256 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 143/154 2026-04-01T13:53:56.257 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:56.257 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-01T13:53:56.257 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:56.336 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 134/154 2026-04-01T13:53:56.351 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 135/154 2026-04-01T13:53:56.405 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 144/154 2026-04-01T13:53:56.431 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 144/154 2026-04-01T13:53:56.431 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:56.431 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-01T13:53:56.431 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-01T13:53:56.431 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-01T13:53:56.431 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:53:56.893 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 134/154 2026-04-01T13:53:56.907 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 135/154 2026-04-01T13:53:56.919 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 135/154 2026-04-01T13:53:56.923 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 136/154 2026-04-01T13:53:56.939 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 136/154 2026-04-01T13:53:56.940 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 137/154 2026-04-01T13:53:57.006 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 137/154 2026-04-01T13:53:57.058 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 138/154 2026-04-01T13:53:57.061 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 139/154 2026-04-01T13:53:57.084 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 139/154 2026-04-01T13:53:57.084 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:57.084 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-01T13:53:57.084 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-01T13:53:57.084 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-01T13:53:57.084 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:57.101 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 140/154 2026-04-01T13:53:57.112 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 140/154 2026-04-01T13:53:57.498 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 135/154 2026-04-01T13:53:57.501 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 136/154 2026-04-01T13:53:57.516 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 136/154 2026-04-01T13:53:57.518 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 137/154 2026-04-01T13:53:57.586 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 137/154 2026-04-01T13:53:57.639 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 138/154 2026-04-01T13:53:57.642 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 139/154 2026-04-01T13:53:57.667 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 139/154 2026-04-01T13:53:57.667 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:57.667 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-01T13:53:57.667 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-01T13:53:57.667 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-01T13:53:57.667 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:57.684 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 140/154 2026-04-01T13:53:57.700 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 140/154 2026-04-01T13:53:58.312 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 141/154 2026-04-01T13:53:58.328 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 142/154 2026-04-01T13:53:58.353 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 142/154 2026-04-01T13:53:58.353 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:58.353 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-01T13:53:58.353 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-01T13:53:58.353 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-01T13:53:58.353 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:58.368 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 143/154 2026-04-01T13:53:58.390 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 143/154 2026-04-01T13:53:58.390 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:58.390 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-01T13:53:58.390 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:58.548 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 144/154 2026-04-01T13:53:58.574 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 144/154 2026-04-01T13:53:58.574 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:58.574 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-01T13:53:58.574 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-01T13:53:58.574 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-01T13:53:58.574 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:53:58.990 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 141/154 2026-04-01T13:53:58.994 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 142/154 2026-04-01T13:53:59.019 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 142/154 2026-04-01T13:53:59.019 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:59.019 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-01T13:53:59.019 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-01T13:53:59.019 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-01T13:53:59.019 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:59.031 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 143/154 2026-04-01T13:53:59.057 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 143/154 2026-04-01T13:53:59.057 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:59.057 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-01T13:53:59.057 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:53:59.218 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 144/154 2026-04-01T13:53:59.247 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 144/154 2026-04-01T13:53:59.247 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T13:53:59.247 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-01T13:53:59.247 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-01T13:53:59.247 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-01T13:53:59.247 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:00.989 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 145/154 2026-04-01T13:54:01.022 INFO:teuthology.orchestra.run.vm09.stdout: Installing : ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 146/154 2026-04-01T13:54:01.029 INFO:teuthology.orchestra.run.vm09.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/154 2026-04-01T13:54:01.037 INFO:teuthology.orchestra.run.vm09.stdout: Installing : libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clys 148/154 2026-04-01T13:54:01.048 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 149/154 2026-04-01T13:54:01.056 INFO:teuthology.orchestra.run.vm09.stdout: Installing : rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 150/154 2026-04-01T13:54:01.076 INFO:teuthology.orchestra.run.vm09.stdout: Installing : bzip2-1.0.8-10.el9_5.x86_64 151/154 2026-04-01T13:54:01.081 INFO:teuthology.orchestra.run.vm09.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 152/154 2026-04-01T13:54:01.081 INFO:teuthology.orchestra.run.vm09.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 153/154 2026-04-01T13:54:01.101 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 153/154 2026-04-01T13:54:01.101 INFO:teuthology.orchestra.run.vm09.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 2/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 3/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 5/154 2026-04-01T13:54:02.732 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 6/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 7/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 8/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 9/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 10/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 11/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 12/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.cly 13/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clys 14/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 15/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 16/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 17/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso 18/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 19/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 20/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el 21/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 22/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso 23/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 24/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 25/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 26/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 27/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 28/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 29/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 30/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 31/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 32/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 33/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 34/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 35/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 36/154 2026-04-01T13:54:02.733 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 37/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 38/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 39/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 40/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 41/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 42/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 43/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 44/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 45/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 46/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 47/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 48/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 49/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 50/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 51/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 52/154 2026-04-01T13:54:02.735 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 53/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 54/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cheroot-10.0.1-5.el9.noarch 55/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cherrypy-18.10.0-5.el9.noarch 56/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 57/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 58/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 59/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 60/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 61/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 62/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 63/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 64/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 65/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 66/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 67/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 68/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 69/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 70/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 71/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 72/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 73/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 74/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 75/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 76/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 77/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 78/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 79/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 80/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 81/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 82/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 83/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : re2-1:20211101-20.el9.x86_64 84/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 85/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 86/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : bzip2-1.0.8-10.el9_5.x86_64 87/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : c-ares-1.19.1-2.el9_4.x86_64 88/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : cryptsetup-2.7.2-4.el9.x86_64 89/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 90/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 91/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 92/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libgfortran-11.5.0-11.el9.x86_64 93/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libquadmath-11.5.0-11.el9.x86_64 94/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lmdb-libs-0.9.29-3.el9.x86_64 95/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 96/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvme-cli-2.13-1.el9.x86_64 97/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 98/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 99/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-cryptography-36.0.1-5.el9_6.x86_64 100/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-ply-3.11-14.el9.0.1.noarch 101/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 102/154 2026-04-01T13:54:02.736 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.0.1.noarch 103/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-2.25.1-10.el9_6.noarch 104/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urllib3-1.26.5-6.el9_7.1.noarch 105/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : smartmontools-1:7.2-9.el9.x86_64 106/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : unzip-6.0-59.el9.x86_64 107/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : zip-3.0-35.el9.x86_64 108/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : boost-program-options-1.75.0-13.el9_7.x86_64 109/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-3.0.4-8.el9.0.1.x86_64 110/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 111/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 112/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 113/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 114/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 115/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 116/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 117/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : libxslt-1.1.34-13.el9_6.x86_64 118/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 119/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-5.4.4-4.el9.x86_64 120/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 121/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 122/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : perl-Benchmark-1.23-481.1.el9_6.noarch 123/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 124/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-3.14.0-17.el9_7.x86_64 125/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 126/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-devel-3.9.23-2.el9.x86_64 127/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jinja2-2.11.3-8.el9_5.noarch 128/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-jmespath-1.0.1-1.el9_7.noarch 129/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 130/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 131/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 132/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-1:1.23.5-2.el9_7.x86_64 133/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 134/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 135/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-protobuf-3.14.0-17.el9_7.noarch 136/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-0.4.8-7.el9_7.noarch 137/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 138/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 139/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 140/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-toml-0.10.2-6.el9.0.1.noarch 141/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-24.09.0-1.el9.x86_64 142/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatlib-service-24.09.0-1.el9.x86_64 143/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 144/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 145/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 146/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 147/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 148/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 149/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : protobuf-compiler-3.14.0-17.el9_7.x86_64 150/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 151/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 152/154 2026-04-01T13:54:02.737 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 153/154 2026-04-01T13:54:02.826 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout:Upgraded: 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: boost-program-options-1.75.0-13.el9_7.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: bzip2-1.0.8-10.el9_5.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: c-ares-1.19.1-2.el9_4.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup-2.7.2-4.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: fuse-2.9.9-17.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data-1.46.7-10.el9.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-9.0.0-15.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libconfig-1.7.2-9.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran-11.5.0-11.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libnbd-1.20.3-4.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: liboath-2.6.12-1.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath-11.5.0-11.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.827 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: libunwind-1.6.2-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: libxslt-1.1.34-13.el9_6.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: lmdb-libs-0.9.29-3.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: lua-5.4.4-4.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli-2.13-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: openblas-0.3.29-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: pciutils-3.7.0-7.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: perl-Benchmark-1.23-481.1.el9_6.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-3.14.0-17.el9_7.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler-3.14.0-17.el9_7.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel-2.9.1-2.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot-10.0.1-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy-18.10.0-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography-36.0.1-5.el9_6.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel-3.9.23-2.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2-2.11.3-8.el9_5.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-jmespath-1.0.1-1.el9_7.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-1:1.23.5-2.el9_7.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging-20.9-5.el9.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply-3.11-14.el9.0.1.noarch 2026-04-01T13:54:02.828 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend-3.1.0-2.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf-3.14.0-17.el9_7.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-0.4.8-7.el9_7.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules-0.4.8-7.el9_7.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser-2.20-6.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing-2.4.7-9.el9.0.1.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-2.25.1-10.el9_6.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes-2.5.1-5.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa-4.9-2.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-saml-1.16.0-1.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml-0.10.2-6.el9.0.1.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3-1.26.5-6.el9_7.1.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-24.09.0-1.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service-24.09.0-1.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: re2-1:20211101-20.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: s3cmd-2.4.0-1.el9.noarch 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools-1:7.2-9.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: thrift-0.15.0-4.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: unzip-6.0-59.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: zip-3.0-35.el9.x86_64 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:02.829 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-04-01T13:54:02.920 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:02.994 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 145/154 2026-04-01T13:54:03.031 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 146/154 2026-04-01T13:54:03.040 INFO:teuthology.orchestra.run.vm08.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/154 2026-04-01T13:54:03.047 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clys 148/154 2026-04-01T13:54:03.059 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 149/154 2026-04-01T13:54:03.067 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 150/154 2026-04-01T13:54:03.086 INFO:teuthology.orchestra.run.vm08.stdout: Installing : bzip2-1.0.8-10.el9_5.x86_64 151/154 2026-04-01T13:54:03.091 INFO:teuthology.orchestra.run.vm08.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 152/154 2026-04-01T13:54:03.091 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 153/154 2026-04-01T13:54:03.113 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 153/154 2026-04-01T13:54:03.113 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:03.880 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 145/154 2026-04-01T13:54:03.971 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 146/154 2026-04-01T13:54:04.088 INFO:teuthology.orchestra.run.vm06.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 147/154 2026-04-01T13:54:04.224 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clys 148/154 2026-04-01T13:54:04.362 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 149/154 2026-04-01T13:54:04.455 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 150/154 2026-04-01T13:54:04.595 INFO:teuthology.orchestra.run.vm06.stdout: Installing : bzip2-1.0.8-10.el9_5.x86_64 151/154 2026-04-01T13:54:04.599 INFO:teuthology.orchestra.run.vm06.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 152/154 2026-04-01T13:54:04.600 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 153/154 2026-04-01T13:54:04.622 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 153/154 2026-04-01T13:54:04.622 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 2/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 3/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 5/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 6/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 7/154 2026-04-01T13:54:04.814 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 8/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 9/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 10/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 11/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 12/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.cly 13/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clys 14/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 15/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 16/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 17/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso 18/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 19/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 20/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el 21/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 22/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso 23/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 24/154 2026-04-01T13:54:04.815 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 25/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 26/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 27/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 28/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 29/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 30/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 31/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 32/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 33/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 34/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 35/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 36/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 37/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 38/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 39/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 40/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 41/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 42/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 43/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 44/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 45/154 2026-04-01T13:54:04.816 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 46/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 47/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 48/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 49/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 50/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 51/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 52/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 53/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 54/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-5.el9.noarch 55/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.10.0-5.el9.noarch 56/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 57/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 58/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 59/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 60/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 61/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 62/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 63/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 64/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 65/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 66/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 67/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 68/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 69/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 70/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 71/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 72/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 73/154 2026-04-01T13:54:04.817 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 74/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 75/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 76/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 77/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 78/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 79/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 80/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 81/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 82/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 83/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 84/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 85/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 86/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : bzip2-1.0.8-10.el9_5.x86_64 87/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : c-ares-1.19.1-2.el9_4.x86_64 88/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.7.2-4.el9.x86_64 89/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 90/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 91/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 92/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-11.el9.x86_64 93/154 2026-04-01T13:54:04.818 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-11.el9.x86_64 94/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lmdb-libs-0.9.29-3.el9.x86_64 95/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 96/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvme-cli-2.13-1.el9.x86_64 97/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 98/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 99/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9_6.x86_64 100/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.0.1.noarch 101/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 102/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.0.1.noarch 103/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9_6.noarch 104/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-6.el9_7.1.noarch 105/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : smartmontools-1:7.2-9.el9.x86_64 106/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 107/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 108/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9_7.x86_64 109/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-8.el9.0.1.x86_64 110/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 111/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 112/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 113/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 114/154 2026-04-01T13:54:04.819 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 115/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 116/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 117/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-13.el9_6.x86_64 118/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 119/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 120/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 121/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 122/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : perl-Benchmark-1.23-481.1.el9_6.noarch 123/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 124/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9_7.x86_64 125/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 126/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.23-2.el9.x86_64 127/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9_5.noarch 128/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jmespath-1.0.1-1.el9_7.noarch 129/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 130/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 131/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 132/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9_7.x86_64 133/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 134/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 135/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9_7.noarch 136/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9_7.noarch 137/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 138/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 139/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 140/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.0.1.noarch 141/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-24.09.0-1.el9.x86_64 142/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-24.09.0-1.el9.x86_64 143/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 144/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 145/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 146/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 147/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 148/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 149/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9_7.x86_64 150/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 151/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 152/154 2026-04-01T13:54:04.820 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 153/154 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout:Upgraded: 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9_7.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: bzip2-1.0.8-10.el9_5.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: c-ares-1.19.1-2.el9_4.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.909 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.7.2-4.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: fuse-2.9.9-17.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-11.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-11.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: libxslt-1.1.34-13.el9_6.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: lmdb-libs-0.9.29-3.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli-2.13-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-04-01T13:54:04.910 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: perl-Benchmark-1.23-481.1.el9_6.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9_7.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9_7.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-5.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.10.0-5.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9_6.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.23-2.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9_5.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath-1.0.1-1.el9_7.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9_7.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.0.1.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9_7.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9_7.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9_7.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-04-01T13:54:04.911 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing-2.4.7-9.el9.0.1.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9_6.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml-1.16.0-1.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.0.1.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-6.el9_7.1.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-24.09.0-1.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-24.09.0-1.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: s3cmd-2.4.0-1.el9.noarch 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools-1:7.2-9.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:04.912 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-04-01T13:54:05.009 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:06.518 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:06.518 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/154 2026-04-01T13:54:06.518 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 2/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 3/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 4/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 5/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 6/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 7/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 8/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 9/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 10/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 11/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 12/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.cly 13/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clys 14/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.cly 15/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 16/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 17/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso 18/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.cly 19/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 20/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el 21/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 22/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso 23/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 24/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 25/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 26/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 27/154 2026-04-01T13:54:06.519 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 28/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 29/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 30/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 31/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 32/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 33/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 34/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 35/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 36/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 37/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 38/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 39/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 40/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 41/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 42/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 43/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 44/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 45/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 46/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 47/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 48/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 49/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 50/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 51/154 2026-04-01T13:54:06.521 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 52/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 53/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 54/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-5.el9.noarch 55/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.10.0-5.el9.noarch 56/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 57/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 58/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 59/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 60/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 61/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 62/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 63/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 64/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 65/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 66/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 67/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 68/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 69/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 70/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 71/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 72/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 73/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 74/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 75/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 76/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 77/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 78/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 79/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 80/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 81/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 82/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 83/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 84/154 2026-04-01T13:54:06.522 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 85/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 86/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : bzip2-1.0.8-10.el9_5.x86_64 87/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : c-ares-1.19.1-2.el9_4.x86_64 88/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.7.2-4.el9.x86_64 89/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 90/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 91/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 92/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-11.el9.x86_64 93/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-11.el9.x86_64 94/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lmdb-libs-0.9.29-3.el9.x86_64 95/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 96/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.13-1.el9.x86_64 97/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 98/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 99/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9_6.x86_64 100/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.0.1.noarch 101/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 102/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.0.1.noarch 103/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9_6.noarch 104/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-6.el9_7.1.noarch 105/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : smartmontools-1:7.2-9.el9.x86_64 106/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 107/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 108/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9_7.x86_64 109/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-8.el9.0.1.x86_64 110/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 111/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 112/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 113/154 2026-04-01T13:54:06.524 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 114/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 115/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 116/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 117/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-13.el9_6.x86_64 118/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 119/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 120/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 121/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 122/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : perl-Benchmark-1.23-481.1.el9_6.noarch 123/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 124/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9_7.x86_64 125/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 126/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.23-2.el9.x86_64 127/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9_5.noarch 128/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9_7.noarch 129/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 130/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 131/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 132/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9_7.x86_64 133/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 134/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 135/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9_7.noarch 136/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9_7.noarch 137/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 138/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 139/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 140/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.0.1.noarch 141/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-24.09.0-1.el9.x86_64 142/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-24.09.0-1.el9.x86_64 143/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 144/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 145/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 146/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 147/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 148/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 149/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9_7.x86_64 150/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 151/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 152/154 2026-04-01T13:54:06.525 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 153/154 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 154/154 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-04-01T13:54:06.617 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9_7.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: bzip2-1.0.8-10.el9_5.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: c-ares-1.19.1-2.el9_4.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:20.2.0-9.ge2ce8426bc5.el9.clyso.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.7.2-4.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: fuse-2.9.9-17.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-proxy2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-04-01T13:54:06.618 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-11.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-11.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-13.el9_6.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: lmdb-libs-0.9.29-3.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.13-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: perl-Benchmark-1.23-481.1.el9_6.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9_7.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9_7.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-5.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.10.0-5.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9_6.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.23-2.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-04-01T13:54:06.619 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9_5.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9_7.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9_7.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.0.1.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9_7.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9_7.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9_7.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.0.1.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9_6.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml-1.16.0-1.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-04-01T13:54:06.620 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.0.1.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-6.el9_7.1.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-24.09.0-1.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-24.09.0-1.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: s3cmd-2.4.0-1.el9.noarch 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools-1:7.2-9.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:06.621 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-04-01T13:54:06.721 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:06.721 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-04-01T13:54:06.721 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-04-01T13:54:06.721 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-04-01T13:54:06.722 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-04-01T13:54:06.722 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:06.722 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-04-01T13:54:06.759 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:06.759 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-04-01T13:54:06.789 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:06.789 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-04-01T13:54:06.823 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-04-01T13:54:06.823 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:06.823 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-04-01T13:54:06.854 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-04-01T13:54:06.922 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:06.922 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/daemon-helper 2026-04-01T13:54:06.949 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-04-01T13:54:07.016 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:07.017 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/daemon-helper 2026-04-01T13:54:07.045 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-04-01T13:54:07.110 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-04-01T13:54:07.111 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:07.111 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-04-01T13:54:07.138 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-04-01T13:54:07.202 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:07.203 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/adjust-ulimits 2026-04-01T13:54:07.228 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-04-01T13:54:07.293 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:07.294 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/adjust-ulimits 2026-04-01T13:54:07.319 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-04-01T13:54:07.386 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-04-01T13:54:07.386 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:07.386 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-04-01T13:54:07.412 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-04-01T13:54:07.478 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:07.478 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/stdin-killer 2026-04-01T13:54:07.503 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-04-01T13:54:07.568 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:07.568 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/usr/bin/stdin-killer 2026-04-01T13:54:07.596 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-04-01T13:54:07.660 INFO:teuthology.run_tasks:Running task ceph... 2026-04-01T13:54:07.700 INFO:tasks.ceph:Making ceph log dir writeable by non-root... 2026-04-01T13:54:07.700 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /var/log/ceph 2026-04-01T13:54:07.702 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /var/log/ceph 2026-04-01T13:54:07.703 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /var/log/ceph 2026-04-01T13:54:07.732 INFO:tasks.ceph:Disabling ceph logrotate... 2026-04-01T13:54:07.732 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/logrotate.d/ceph 2026-04-01T13:54:07.767 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/logrotate.d/ceph 2026-04-01T13:54:07.769 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/logrotate.d/ceph 2026-04-01T13:54:07.800 INFO:tasks.ceph:Creating extra log directories... 2026-04-01T13:54:07.800 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-04-01T13:54:07.831 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-04-01T13:54:07.834 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-04-01T13:54:07.872 INFO:tasks.ceph:Creating ceph cluster ceph... 2026-04-01T13:54:07.872 INFO:tasks.ceph:config {'conf': {'client': {'debug rgw': 20, 'debug rgw dedup': 20, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'global': {'osd_max_pg_log_entries': 10, 'osd_min_pg_log_entries': 10}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'bdev async discard': True, 'bdev enable discard': True, 'bluestore allocator': 'bitmap', 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug ms': 1, 'debug osd': 20, 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd mclock iops capacity threshold hdd': 49000, 'osd objectstore': 'bluestore', 'osd shutdown pgref assert': True}}, 'fs': 'xfs', 'mkfs_options': None, 'mount_options': None, 'skip_mgr_daemons': False, 'log_ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'cpu_profile': set(), 'cluster': 'ceph', 'mon_bind_msgr2': True, 'mon_bind_addrvec': True} 2026-04-01T13:54:07.872 INFO:tasks.ceph:ctx.config {'archive_path': '/archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802', 'branch': 'wip-sse-s3-on-v20.2.0', 'description': 'rgw/dedup/{beast bluestore-bitmap fixed-3-rgw ignore-pg-availability overrides supported-distros/{rocky_latest} tasks/{0-install test_dedup}}', 'email': None, 'first_in_suite': False, 'flavor': 'default', 'job_id': '4802', 'last_in_suite': False, 'machine_type': 'vps', 'name': 'supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps', 'no_nested_subset': False, 'openstack': [{'volumes': {'count': 4, 'size': 10}}], 'os_type': 'rocky', 'os_version': '9.7', 'overrides': {'admin_socket': {'branch': 'wip-sse-s3-on-v20.2.0'}, 'ansible.cephlab': {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}}, 'ceph': {'conf': {'client': {'debug rgw': 20, 'debug rgw dedup': 20, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'global': {'osd_max_pg_log_entries': 10, 'osd_min_pg_log_entries': 10}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'bdev async discard': True, 'bdev enable discard': True, 'bluestore allocator': 'bitmap', 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug ms': 1, 'debug osd': 20, 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd mclock iops capacity threshold hdd': 49000, 'osd objectstore': 'bluestore', 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'fs': 'xfs', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6'}, 'ceph-deploy': {'bluestore': True, 'conf': {'client': {'log file': '/var/log/ceph/ceph-$name.$pid.log'}, 'mon': {}, 'osd': {'bdev async discard': True, 'bdev enable discard': True, 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd objectstore': 'bluestore'}}, 'fs': 'xfs'}, 'cephadm': {'cephadm_binary_url': 'https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:sse-s3-kmip-preview-not-for-production-2'}}, 'install': {'ceph': {'flavor': 'default', 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-9-ge2ce8426bc5/el9.clyso/x86_64'}]}, 'rgw': {'frontend': 'beast', 'storage classes': {'FROZEN': None, 'LUKEWARM': None}}, 's3tests': {'sha1': 'e0c4ff71baef6d5126a0201df5fe54196d89b296'}, 'selinux': {'allowlist': ['scontext=system_u:system_r:getty_t:s0']}, 'thrashosds': {'bdev_inject_crash': 2, 'bdev_inject_crash_probability': 0.5}, 'workunit': {'branch': 'tt-20.2.0-sse-s3-kmip-preview-not-for-production-2', 'sha1': '0d46c98b681455ef9cd34c2d3c6981da4e9f1aea'}}, 'owner': 'supriti', 'priority': 1000, 'repo': 'https://github.com/ceph/ceph.git', 'roles': [['mon.a', 'mon.c', 'mgr.y', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'], ['mon.b', 'mgr.x', 'osd.4', 'osd.5', 'osd.6', 'osd.7', 'client.1'], ['client.2']], 'seed': 343, 'sha1': 'e2ce8426bc5387b913d1b06bd31280ced108bbb6', 'sleep_before_teardown': 0, 'suite': 'rgw', 'suite_branch': 'tt-20.2.0-sse-s3-kmip-preview-not-for-production-2', 'suite_path': '/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa', 'suite_relpath': 'qa', 'suite_repo': 'http://git.local/ceph.git', 'suite_sha1': '0d46c98b681455ef9cd34c2d3c6981da4e9f1aea', 'targets': {'vm06.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBNTOLJDVW6KK/6J5BdWj48jn2txkC0z7tw+YqaQ/6nXMOyhu/56c5vRw26fnzDUT+rhL+2WOGbcS6a3VQB//jnI=', 'vm08.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBAR1cnfhNFk7xCg/QzMr9TAXWR2qIkMIewcDvjVdOdAwsgIREaYdKXhUYOz+HPynWcoyzk33kjDDD7Q8PRbkzjE=', 'vm09.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPVr7OG9quN7TJQPoyWRtARk297W6aw+/g+hvz++NgJFauXwO5rB98MdTAsBBRadXZsdskUFGesjgEbhYM319m0='}, 'tasks': [{'internal.save_config': None}, {'internal.check_lock': None}, {'internal.add_remotes': None}, {'console_log': None}, {'internal.connect': None}, {'internal.push_inventory': None}, {'internal.serialize_remote_roles': None}, {'internal.check_conflict': None}, {'internal.check_ceph_data': None}, {'internal.vm_setup': None}, {'internal.base': None}, {'internal.archive_upload': None}, {'internal.archive': None}, {'internal.coredump': None}, {'internal.sudo': None}, {'internal.syslog': None}, {'internal.timer': None}, {'pcp': None}, {'selinux': None}, {'ansible.cephlab': None}, {'clock': None}, {'install': None}, {'ceph': None}, {'openssl_keys': None}, {'rgw': ['client.0', 'client.1', 'client.2']}, {'tox': ['client.0']}, {'tox': ['client.0']}, {'dedup-tests': {'client.0': {'rgw_server': 'client.0'}}}], 'teuthology': {'fragments_dropped': [], 'meta': {}, 'postmerge': []}, 'teuthology_branch': 'uv2', 'teuthology_repo': 'https://github.com/kshtsk/teuthology', 'teuthology_sha1': 'a59626679648f962bca99d20d35578f2998c8f37', 'timestamp': '2026-04-01_13:45:16', 'tube': 'vps', 'user': 'supriti', 'verbose': False, 'worker_log': '/home/teuthos/.teuthology/dispatcher/dispatcher.vps.1399776'} 2026-04-01T13:54:07.872 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/ceph.data 2026-04-01T13:54:07.896 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/ceph.data 2026-04-01T13:54:07.900 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/ceph.data 2026-04-01T13:54:07.926 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/run/ceph 2026-04-01T13:54:07.952 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/run/ceph 2026-04-01T13:54:07.959 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m0777 -- /var/run/ceph 2026-04-01T13:54:07.996 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:07.996 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-04-01T13:54:08.032 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-04-01T13:54:08.032 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_1 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1051 Links: 1 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-01 13:54:05.062396738 +0000 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-01 13:53:42.304383192 +0000 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-01 13:53:42.304383192 +0000 2026-04-01T13:54:08.089 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-01 13:53:42.304383192 +0000 2026-04-01T13:54:08.089 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-04-01T13:54:08.156 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-01T13:54:08.156 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-01T13:54:08.156 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000229599 s, 2.2 MB/s 2026-04-01T13:54:08.157 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-04-01T13:54:08.217 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_2 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1061 Links: 1 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-01 13:54:05.062396738 +0000 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-01 13:53:42.305383193 +0000 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-01 13:53:42.305383193 +0000 2026-04-01T13:54:08.276 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-01 13:53:42.305383193 +0000 2026-04-01T13:54:08.276 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-04-01T13:54:08.343 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-01T13:54:08.343 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-01T13:54:08.343 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000196588 s, 2.6 MB/s 2026-04-01T13:54:08.344 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-04-01T13:54:08.405 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_3 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1056 Links: 1 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-01 13:54:05.063396739 +0000 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-01 13:53:42.304383192 +0000 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-01 13:53:42.304383192 +0000 2026-04-01T13:54:08.462 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-01 13:53:42.304383192 +0000 2026-04-01T13:54:08.463 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-04-01T13:54:08.528 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-01T13:54:08.528 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-01T13:54:08.528 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000212559 s, 2.4 MB/s 2026-04-01T13:54:08.529 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-04-01T13:54:08.587 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_4 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1068 Links: 1 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-01 13:54:05.063396739 +0000 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-01 13:53:42.308383194 +0000 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-01 13:53:42.308383194 +0000 2026-04-01T13:54:08.646 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-01 13:53:42.308383194 +0000 2026-04-01T13:54:08.646 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-04-01T13:54:08.711 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-01T13:54:08.711 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-01T13:54:08.711 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000142636 s, 3.6 MB/s 2026-04-01T13:54:08.712 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-04-01T13:54:08.771 INFO:tasks.ceph:osd dev map: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:08.771 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:08.771 DEBUG:teuthology.orchestra.run.vm08:> dd if=/scratch_devs of=/dev/stdout 2026-04-01T13:54:08.790 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-04-01T13:54:08.790 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_1 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 1063 Links: 1 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-04-01 13:54:03.383548858 +0000 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-04-01 13:53:41.854584051 +0000 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-04-01 13:53:41.854584051 +0000 2026-04-01T13:54:08.848 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-04-01 13:53:41.854584051 +0000 2026-04-01T13:54:08.848 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-04-01T13:54:08.916 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-04-01T13:54:08.916 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-04-01T13:54:08.916 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000183534 s, 2.8 MB/s 2026-04-01T13:54:08.917 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-04-01T13:54:08.977 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_2 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 1059 Links: 1 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-04-01 13:54:03.383548858 +0000 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-04-01 13:53:41.850584070 +0000 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-04-01 13:53:41.850584070 +0000 2026-04-01T13:54:09.035 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-04-01 13:53:41.850584070 +0000 2026-04-01T13:54:09.035 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-04-01T13:54:09.099 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-04-01T13:54:09.100 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-04-01T13:54:09.100 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000147556 s, 3.5 MB/s 2026-04-01T13:54:09.101 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-04-01T13:54:09.159 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_3 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 1051 Links: 1 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-04-01 13:54:03.383548858 +0000 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-04-01 13:53:41.847584083 +0000 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-04-01 13:53:41.847584083 +0000 2026-04-01T13:54:09.218 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-04-01 13:53:41.847584083 +0000 2026-04-01T13:54:09.218 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-04-01T13:54:09.282 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-04-01T13:54:09.282 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-04-01T13:54:09.282 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000145623 s, 3.5 MB/s 2026-04-01T13:54:09.283 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-04-01T13:54:09.341 DEBUG:teuthology.orchestra.run.vm08:> stat /dev/vg_nvme/lv_4 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout:Device: 5h/5d Inode: 1070 Links: 1 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout:Context: system_u:object_r:device_t:s0 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout:Access: 2026-04-01 13:54:03.383548858 +0000 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout:Modify: 2026-04-01 13:53:41.855584046 +0000 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout:Change: 2026-04-01 13:53:41.855584046 +0000 2026-04-01T13:54:09.398 INFO:teuthology.orchestra.run.vm08.stdout: Birth: 2026-04-01 13:53:41.855584046 +0000 2026-04-01T13:54:09.398 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-04-01T13:54:09.463 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records in 2026-04-01T13:54:09.463 INFO:teuthology.orchestra.run.vm08.stderr:1+0 records out 2026-04-01T13:54:09.463 INFO:teuthology.orchestra.run.vm08.stderr:512 bytes copied, 0.000132898 s, 3.9 MB/s 2026-04-01T13:54:09.465 DEBUG:teuthology.orchestra.run.vm08:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-04-01T13:54:09.521 INFO:tasks.ceph:osd dev map: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:09.521 INFO:tasks.ceph:remote_to_roles_to_devs: {Remote(name='ubuntu@vm06.local'): {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'}, Remote(name='ubuntu@vm08.local'): {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'}} 2026-04-01T13:54:09.521 INFO:tasks.ceph:Generating config... 2026-04-01T13:54:09.522 INFO:tasks.ceph:[client] debug rgw = 20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[client] debug rgw dedup = 20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[client] setgroup = ceph 2026-04-01T13:54:09.522 INFO:tasks.ceph:[client] setuser = ceph 2026-04-01T13:54:09.522 INFO:tasks.ceph:[global] osd_max_pg_log_entries = 10 2026-04-01T13:54:09.522 INFO:tasks.ceph:[global] osd_min_pg_log_entries = 10 2026-04-01T13:54:09.522 INFO:tasks.ceph:[mgr] debug mgr = 20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[mgr] debug ms = 1 2026-04-01T13:54:09.522 INFO:tasks.ceph:[mon] debug mon = 20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[mon] debug ms = 1 2026-04-01T13:54:09.522 INFO:tasks.ceph:[mon] debug paxos = 20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] bdev async discard = True 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] bdev enable discard = True 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] bluestore allocator = bitmap 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] bluestore block size = 96636764160 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] bluestore fsck on mount = True 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] debug bluefs = 1/20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] debug bluestore = 1/20 2026-04-01T13:54:09.522 INFO:tasks.ceph:[osd] debug ms = 1 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] debug osd = 20 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] debug rocksdb = 4/10 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] mon osd backfillfull_ratio = 0.85 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] mon osd full ratio = 0.9 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] mon osd nearfull ratio = 0.8 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] osd failsafe full ratio = 0.95 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] osd mclock iops capacity threshold hdd = 49000 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] osd objectstore = bluestore 2026-04-01T13:54:09.523 INFO:tasks.ceph:[osd] osd shutdown pgref assert = True 2026-04-01T13:54:09.523 INFO:tasks.ceph:Setting up mon.a... 2026-04-01T13:54:09.523 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring /etc/ceph/ceph.keyring 2026-04-01T13:54:09.559 INFO:teuthology.orchestra.run.vm06.stdout:creating /etc/ceph/ceph.keyring 2026-04-01T13:54:09.561 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=mon. /etc/ceph/ceph.keyring 2026-04-01T13:54:09.642 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-01T13:54:09.708 DEBUG:tasks.ceph:Ceph mon addresses: [('mon.a', '192.168.123.106'), ('mon.c', '[v2:192.168.123.106:3301,v1:192.168.123.106:6790]'), ('mon.b', '192.168.123.108')] 2026-04-01T13:54:09.708 DEBUG:tasks.ceph:writing out conf {'global': {'chdir': '', 'pid file': '/var/run/ceph/$cluster-$name.pid', 'auth supported': 'cephx', 'filestore xattr use omap': 'true', 'mon clock drift allowed': '1.000', 'osd crush chooseleaf type': '0', 'auth debug': 'true', 'ms die on old message': 'true', 'ms die on bug': 'true', 'mon max pg per osd': '10000', 'mon pg warn max object skew': '0', 'osd_pool_default_pg_autoscale_mode': 'off', 'osd pool default size': '2', 'mon osd allow primary affinity': 'true', 'mon osd allow pg remap': 'true', 'mon warn on legacy crush tunables': 'false', 'mon warn on crush straw calc version zero': 'false', 'mon warn on no sortbitwise': 'false', 'mon warn on osd down out interval zero': 'false', 'mon warn on too few osds': 'false', 'mon_warn_on_pool_pg_num_not_power_of_two': 'false', 'mon_warn_on_pool_no_redundancy': 'false', 'mon_allow_pool_size_one': 'true', 'osd pool default erasure code profile': 'plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd', 'osd default data pool replay window': '5', 'mon allow pool delete': 'true', 'mon cluster log file level': 'debug', 'debug asserts on shutdown': 'true', 'mon health detail to clog': 'false', 'mon host': '192.168.123.106,[v2:192.168.123.106:3301,v1:192.168.123.106:6790],192.168.123.108', 'osd_max_pg_log_entries': 10, 'osd_min_pg_log_entries': 10}, 'osd': {'osd journal size': '100', 'osd scrub load threshold': '5.0', 'osd scrub max interval': '600', 'osd mclock profile': 'high_recovery_ops', 'osd recover clone overlap': 'true', 'osd recovery max chunk': '1048576', 'osd debug shutdown': 'true', 'osd debug op order': 'true', 'osd debug verify stray on activate': 'true', 'osd debug trim objects': 'true', 'osd open classes on start': 'true', 'osd debug pg log writeout': 'true', 'osd deep scrub update digest min age': '30', 'osd map max advance': '10', 'journal zero on create': 'true', 'filestore ondisk finisher threads': '3', 'filestore apply finisher threads': '3', 'bdev debug aio': 'true', 'osd debug misdirected ops': 'true', 'bdev async discard': True, 'bdev enable discard': True, 'bluestore allocator': 'bitmap', 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug ms': 1, 'debug osd': 20, 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd mclock iops capacity threshold hdd': 49000, 'osd objectstore': 'bluestore', 'osd shutdown pgref assert': True}, 'mgr': {'debug ms': 1, 'debug mgr': 20, 'debug mon': '20', 'debug auth': '20', 'mon reweight min pgs per osd': '4', 'mon reweight min bytes per osd': '10', 'mgr/telemetry/nag': 'false'}, 'mon': {'debug ms': 1, 'debug mon': 20, 'debug paxos': 20, 'debug auth': '20', 'mon data avail warn': '5', 'mon mgr mkfs grace': '240', 'mon reweight min pgs per osd': '4', 'mon osd reporter subtree level': 'osd', 'mon osd prime pg temp': 'true', 'mon reweight min bytes per osd': '10', 'auth mon ticket ttl': '660', 'auth service ticket ttl': '240', 'mon_warn_on_insecure_global_id_reclaim': 'false', 'mon_warn_on_insecure_global_id_reclaim_allowed': 'false', 'mon_down_mkfs_grace': '2m', 'mon_warn_on_filestore_osds': 'false'}, 'client': {'rgw cache enabled': 'true', 'rgw enable ops log': 'true', 'rgw enable usage log': 'true', 'log file': '/var/log/ceph/$cluster-$name.$pid.log', 'admin socket': '/var/run/ceph/$cluster-$name.$pid.asok', 'debug rgw': 20, 'debug rgw dedup': 20, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'mon.a': {}, 'mon.c': {}, 'mon.b': {}} 2026-04-01T13:54:09.708 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:09.708 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/ceph.tmp.conf 2026-04-01T13:54:09.765 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage monmaptool -c /home/ubuntu/cephtest/ceph.tmp.conf --create --clobber --enable-all-features --add a 192.168.123.106 --addv c '[v2:192.168.123.106:3301,v1:192.168.123.106:6790]' --add b 192.168.123.108 --print /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool: monmap file /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool: generated fsid 8a8efea3-ddf7-4150-909c-57db5b081e35 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:setting min_mon_release = tentacle 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:epoch 0 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:fsid 8a8efea3-ddf7-4150-909c-57db5b081e35 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:last_changed 2026-04-01T13:54:09.844494+0000 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:created 2026-04-01T13:54:09.844494+0000 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:min_mon_release 20 (tentacle) 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:election_strategy: 1 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.a 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:1: [v2:192.168.123.108:3300/0,v1:192.168.123.108:6789/0] mon.b 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:2: [v2:192.168.123.106:3301/0,v1:192.168.123.106:6790/0] mon.c 2026-04-01T13:54:09.843 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool: writing epoch 0 to /home/ubuntu/cephtest/ceph.monmap (3 monitors) 2026-04-01T13:54:09.845 DEBUG:teuthology.orchestra.run.vm06:> rm -- /home/ubuntu/cephtest/ceph.tmp.conf 2026-04-01T13:54:09.905 INFO:tasks.ceph:Writing /etc/ceph/ceph.conf for FSID 8a8efea3-ddf7-4150-909c-57db5b081e35... 2026-04-01T13:54:09.905 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/ceph.conf && sudo chmod 0644 /etc/ceph/ceph.conf > /dev/null 2026-04-01T13:54:09.946 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/ceph.conf && sudo chmod 0644 /etc/ceph/ceph.conf > /dev/null 2026-04-01T13:54:09.947 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/ceph.conf && sudo chmod 0644 /etc/ceph/ceph.conf > /dev/null 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: chdir = "" 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: auth supported = cephx 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: filestore xattr use omap = true 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: mon clock drift allowed = 1.000 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: osd crush chooseleaf type = 0 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: auth debug = true 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.988 INFO:teuthology.orchestra.run.vm06.stdout: ms die on old message = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: ms die on bug = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon max pg per osd = 10000 # >= luminous 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon pg warn max object skew = 0 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: # disable pg_autoscaler by default for new pools 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd_pool_default_pg_autoscale_mode = off 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default size = 2 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow primary affinity = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow pg remap = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on legacy crush tunables = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on crush straw calc version zero = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on no sortbitwise = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on osd down out interval zero = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on too few osds = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_no_redundancy = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon_allow_pool_size_one = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd default data pool replay window = 5 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon allow pool delete = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon cluster log file level = debug 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: debug asserts on shutdown = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon health detail to clog = false 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: mon host = "192.168.123.106,[v2:192.168.123.106:3301,v1:192.168.123.106:6790],192.168.123.108" 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_pg_log_entries = 10 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd_min_pg_log_entries = 10 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: fsid = 8a8efea3-ddf7-4150-909c-57db5b081e35 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout:[osd] 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd journal size = 100 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub load threshold = 5.0 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub max interval = 600 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock profile = high_recovery_ops 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd recover clone overlap = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd recovery max chunk = 1048576 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd debug shutdown = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd debug op order = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd debug verify stray on activate = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd debug trim objects = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd open classes on start = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd debug pg log writeout = true 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd deep scrub update digest min age = 30 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: osd map max advance = 10 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.989 INFO:teuthology.orchestra.run.vm06.stdout: journal zero on create = true 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: filestore ondisk finisher threads = 3 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: filestore apply finisher threads = 3 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: bdev debug aio = true 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: osd debug misdirected ops = true 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: bdev async discard = True 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: bdev enable discard = True 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: bluestore allocator = bitmap 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: bluestore block size = 96636764160 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: bluestore fsck on mount = True 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug bluefs = 1/20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug bluestore = 1/20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug osd = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug rocksdb = 4/10 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon osd backfillfull_ratio = 0.85 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon osd full ratio = 0.9 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon osd nearfull ratio = 0.8 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: osd failsafe full ratio = 0.95 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock iops capacity threshold hdd = 49000 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: osd objectstore = bluestore 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: osd shutdown pgref assert = True 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout:[mgr] 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug mgr = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mgr/telemetry/nag = false 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout:[mon] 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug paxos = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon data avail warn = 5 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon mgr mkfs grace = 240 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon osd reporter subtree level = osd 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon osd prime pg temp = true 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: auth mon ticket ttl = 660 # 11m 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: auth service ticket ttl = 240 # 4m 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: # don't complain about insecure global_id in the test suite 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: # 1m isn't quite enough 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon_down_mkfs_grace = 2m 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_filestore_osds = false 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout:[client] 2026-04-01T13:54:09.990 INFO:teuthology.orchestra.run.vm06.stdout: rgw cache enabled = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable ops log = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable usage log = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw = 20 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw dedup = 20 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: setgroup = ceph 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout: setuser = ceph 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout:[mon.a] 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout:[mon.c] 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm06.stdout:[mon.b] 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout:[global] 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: chdir = "" 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: auth supported = cephx 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: filestore xattr use omap = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon clock drift allowed = 1.000 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: osd crush chooseleaf type = 0 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: auth debug = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: ms die on old message = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: ms die on bug = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon max pg per osd = 10000 # >= luminous 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon pg warn max object skew = 0 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: # disable pg_autoscaler by default for new pools 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: osd_pool_default_pg_autoscale_mode = off 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: osd pool default size = 2 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon osd allow primary affinity = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon osd allow pg remap = true 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon warn on legacy crush tunables = false 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon warn on crush straw calc version zero = false 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon warn on no sortbitwise = false 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon warn on osd down out interval zero = false 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon warn on too few osds = false 2026-04-01T13:54:09.991 INFO:teuthology.orchestra.run.vm09.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: mon_warn_on_pool_no_redundancy = false 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: mon_allow_pool_size_one = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd default data pool replay window = 5 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: mon allow pool delete = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: mon cluster log file level = debug 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: debug asserts on shutdown = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: mon health detail to clog = false 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: mon host = "192.168.123.106,[v2:192.168.123.106:3301,v1:192.168.123.106:6790],192.168.123.108" 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd_max_pg_log_entries = 10 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd_min_pg_log_entries = 10 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: fsid = 8a8efea3-ddf7-4150-909c-57db5b081e35 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout:[osd] 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd journal size = 100 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd scrub load threshold = 5.0 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd scrub max interval = 600 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd mclock profile = high_recovery_ops 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd recover clone overlap = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd recovery max chunk = 1048576 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd debug shutdown = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd debug op order = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd debug verify stray on activate = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd debug trim objects = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd open classes on start = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd debug pg log writeout = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd deep scrub update digest min age = 30 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd map max advance = 10 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: journal zero on create = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: filestore ondisk finisher threads = 3 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: filestore apply finisher threads = 3 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: bdev debug aio = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: osd debug misdirected ops = true 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: bdev async discard = True 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: bdev enable discard = True 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: bluestore allocator = bitmap 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: bluestore block size = 96636764160 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: bluestore fsck on mount = True 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: debug bluefs = 1/20 2026-04-01T13:54:09.992 INFO:teuthology.orchestra.run.vm09.stdout: debug bluestore = 1/20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug ms = 1 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug osd = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug rocksdb = 4/10 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon osd backfillfull_ratio = 0.85 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon osd full ratio = 0.9 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon osd nearfull ratio = 0.8 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: osd failsafe full ratio = 0.95 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: osd mclock iops capacity threshold hdd = 49000 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: osd objectstore = bluestore 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: osd shutdown pgref assert = True 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout:[mgr] 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug ms = 1 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug mgr = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug mon = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug auth = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon reweight min pgs per osd = 4 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon reweight min bytes per osd = 10 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mgr/telemetry/nag = false 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout:[mon] 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug ms = 1 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug mon = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug paxos = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: debug auth = 20 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon data avail warn = 5 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon mgr mkfs grace = 240 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon reweight min pgs per osd = 4 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon osd reporter subtree level = osd 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon osd prime pg temp = true 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon reweight min bytes per osd = 10 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: auth mon ticket ttl = 660 # 11m 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: auth service ticket ttl = 240 # 4m 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: # don't complain about insecure global_id in the test suite 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: # 1m isn't quite enough 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon_down_mkfs_grace = 2m 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: mon_warn_on_filestore_osds = false 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout:[client] 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: rgw cache enabled = true 2026-04-01T13:54:09.993 INFO:teuthology.orchestra.run.vm09.stdout: rgw enable ops log = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: rgw enable usage log = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: debug rgw = 20 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: debug rgw dedup = 20 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: setgroup = ceph 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout: setuser = ceph 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout:[mon.a] 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout:[mon.c] 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm09.stdout:[mon.b] 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout:[global] 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: chdir = "" 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: auth supported = cephx 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: filestore xattr use omap = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon clock drift allowed = 1.000 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: osd crush chooseleaf type = 0 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: auth debug = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: ms die on old message = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: ms die on bug = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon max pg per osd = 10000 # >= luminous 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon pg warn max object skew = 0 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: # disable pg_autoscaler by default for new pools 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: osd_pool_default_pg_autoscale_mode = off 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default size = 2 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow primary affinity = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow pg remap = true 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on legacy crush tunables = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on crush straw calc version zero = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on no sortbitwise = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on osd down out interval zero = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on too few osds = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_no_redundancy = false 2026-04-01T13:54:09.994 INFO:teuthology.orchestra.run.vm08.stdout: mon_allow_pool_size_one = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd default data pool replay window = 5 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon allow pool delete = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon cluster log file level = debug 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: debug asserts on shutdown = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon health detail to clog = false 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon host = "192.168.123.106,[v2:192.168.123.106:3301,v1:192.168.123.106:6790],192.168.123.108" 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_pg_log_entries = 10 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd_min_pg_log_entries = 10 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: fsid = 8a8efea3-ddf7-4150-909c-57db5b081e35 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout:[osd] 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd journal size = 100 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub load threshold = 5.0 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub max interval = 600 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock profile = high_recovery_ops 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd recover clone overlap = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd recovery max chunk = 1048576 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd debug shutdown = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd debug op order = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd debug verify stray on activate = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd debug trim objects = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd open classes on start = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd debug pg log writeout = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd deep scrub update digest min age = 30 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd map max advance = 10 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: journal zero on create = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: filestore ondisk finisher threads = 3 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: filestore apply finisher threads = 3 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: bdev debug aio = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: osd debug misdirected ops = true 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: bdev async discard = True 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: bdev enable discard = True 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: bluestore allocator = bitmap 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: bluestore block size = 96636764160 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: bluestore fsck on mount = True 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: debug bluefs = 1/20 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: debug bluestore = 1/20 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: debug osd = 20 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: debug rocksdb = 4/10 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon osd backfillfull_ratio = 0.85 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon osd full ratio = 0.9 2026-04-01T13:54:09.995 INFO:teuthology.orchestra.run.vm08.stdout: mon osd nearfull ratio = 0.8 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: osd failsafe full ratio = 0.95 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock iops capacity threshold hdd = 49000 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: osd objectstore = bluestore 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: osd shutdown pgref assert = True 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout:[mgr] 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug mgr = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mgr/telemetry/nag = false 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout:[mon] 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug paxos = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon data avail warn = 5 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon mgr mkfs grace = 240 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon osd reporter subtree level = osd 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon osd prime pg temp = true 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: auth mon ticket ttl = 660 # 11m 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: auth service ticket ttl = 240 # 4m 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: # don't complain about insecure global_id in the test suite 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: # 1m isn't quite enough 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon_down_mkfs_grace = 2m 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_filestore_osds = false 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout:[client] 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: rgw cache enabled = true 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable ops log = true 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable usage log = true 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw = 20 2026-04-01T13:54:09.996 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw dedup = 20 2026-04-01T13:54:09.997 INFO:teuthology.orchestra.run.vm08.stdout: setgroup = ceph 2026-04-01T13:54:09.997 INFO:teuthology.orchestra.run.vm08.stdout: setuser = ceph 2026-04-01T13:54:09.997 INFO:teuthology.orchestra.run.vm08.stdout:[mon.a] 2026-04-01T13:54:09.997 INFO:teuthology.orchestra.run.vm08.stdout:[mon.c] 2026-04-01T13:54:09.997 INFO:teuthology.orchestra.run.vm08.stdout:[mon.b] 2026-04-01T13:54:10.000 INFO:tasks.ceph:Creating admin key on mon.a... 2026-04-01T13:54:10.000 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=client.admin --cap mon 'allow *' --cap osd 'allow *' --cap mds 'allow *' --cap mgr 'allow *' /etc/ceph/ceph.keyring 2026-04-01T13:54:10.081 INFO:tasks.ceph:Copying monmap to all nodes... 2026-04-01T13:54:10.082 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:10.082 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.keyring of=/dev/stdout 2026-04-01T13:54:10.138 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:10.138 DEBUG:teuthology.orchestra.run.vm06:> dd if=/home/ubuntu/cephtest/ceph.monmap of=/dev/stdout 2026-04-01T13:54:10.193 INFO:tasks.ceph:Sending monmap to node ubuntu@vm06.local 2026-04-01T13:54:10.193 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:10.193 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.keyring 2026-04-01T13:54:10.193 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-01T13:54:10.268 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:10.268 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:10.326 INFO:tasks.ceph:Sending monmap to node ubuntu@vm08.local 2026-04-01T13:54:10.326 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:10.326 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.keyring 2026-04-01T13:54:10.326 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-01T13:54:10.361 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:10.361 DEBUG:teuthology.orchestra.run.vm08:> dd of=/home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:10.419 INFO:tasks.ceph:Sending monmap to node ubuntu@vm09.local 2026-04-01T13:54:10.419 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:10.419 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.keyring 2026-04-01T13:54:10.419 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-01T13:54:10.450 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:10.450 DEBUG:teuthology.orchestra.run.vm09:> dd of=/home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:10.505 INFO:tasks.ceph:Setting up mon nodes... 2026-04-01T13:54:10.505 INFO:tasks.ceph:Setting up mgr nodes... 2026-04-01T13:54:10.505 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/mgr/ceph-y && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=mgr.y /var/lib/ceph/mgr/ceph-y/keyring 2026-04-01T13:54:10.550 INFO:teuthology.orchestra.run.vm06.stdout:creating /var/lib/ceph/mgr/ceph-y/keyring 2026-04-01T13:54:10.553 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/mgr/ceph-x && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=mgr.x /var/lib/ceph/mgr/ceph-x/keyring 2026-04-01T13:54:10.601 INFO:teuthology.orchestra.run.vm08.stdout:creating /var/lib/ceph/mgr/ceph-x/keyring 2026-04-01T13:54:10.603 INFO:tasks.ceph:Setting up mds nodes... 2026-04-01T13:54:10.604 INFO:tasks.ceph_client:Setting up client nodes... 2026-04-01T13:54:10.604 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.0 /etc/ceph/ceph.client.0.keyring && sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-04-01T13:54:10.642 INFO:teuthology.orchestra.run.vm06.stdout:creating /etc/ceph/ceph.client.0.keyring 2026-04-01T13:54:10.653 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.1 /etc/ceph/ceph.client.1.keyring && sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-04-01T13:54:10.691 INFO:teuthology.orchestra.run.vm08.stdout:creating /etc/ceph/ceph.client.1.keyring 2026-04-01T13:54:10.702 DEBUG:teuthology.orchestra.run.vm09:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.2 /etc/ceph/ceph.client.2.keyring && sudo chmod 0644 /etc/ceph/ceph.client.2.keyring 2026-04-01T13:54:10.737 INFO:teuthology.orchestra.run.vm09.stdout:creating /etc/ceph/ceph.client.2.keyring 2026-04-01T13:54:10.747 INFO:tasks.ceph:Running mkfs on osd nodes... 2026-04-01T13:54:10.747 INFO:tasks.ceph:ctx.disk_config.remote_to_roles_to_dev: {Remote(name='ubuntu@vm06.local'): {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'}, Remote(name='ubuntu@vm08.local'): {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'}} 2026-04-01T13:54:10.747 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-0 2026-04-01T13:54:10.773 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:10.773 INFO:tasks.ceph:role: osd.0 2026-04-01T13:54:10.773 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm06.local 2026-04-01T13:54:10.773 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:10.843 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:10.847 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-01T13:54:10.849 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm06.local -o noatime 2026-04-01T13:54:10.849 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/ceph-0 2026-04-01T13:54:10.920 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-0 2026-04-01T13:54:10.989 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-1 2026-04-01T13:54:11.054 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:11.055 INFO:tasks.ceph:role: osd.1 2026-04-01T13:54:11.055 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm06.local 2026-04-01T13:54:11.055 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-04-01T13:54:11.121 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:11.121 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:11.121 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:11.121 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:11.121 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:11.121 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:11.122 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:11.122 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:11.122 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:11.122 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:11.126 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-01T13:54:11.128 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm06.local -o noatime 2026-04-01T13:54:11.128 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/ceph-1 2026-04-01T13:54:11.201 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-1 2026-04-01T13:54:11.270 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-2 2026-04-01T13:54:11.336 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:11.336 INFO:tasks.ceph:role: osd.2 2026-04-01T13:54:11.336 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm06.local 2026-04-01T13:54:11.336 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:11.399 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:11.403 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-01T13:54:11.406 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm06.local -o noatime 2026-04-01T13:54:11.406 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/ceph-2 2026-04-01T13:54:11.473 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-2 2026-04-01T13:54:11.538 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-3 2026-04-01T13:54:11.601 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:11.601 INFO:tasks.ceph:role: osd.3 2026-04-01T13:54:11.601 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_4 on ubuntu@vm06.local 2026-04-01T13:54:11.601 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_4 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_4 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:11.666 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:11.670 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-01T13:54:11.672 INFO:tasks.ceph:mount /dev/vg_nvme/lv_4 on ubuntu@vm06.local -o noatime 2026-04-01T13:54:11.672 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_4 /var/lib/ceph/osd/ceph-3 2026-04-01T13:54:11.740 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-3 2026-04-01T13:54:11.810 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:11.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:11.889+0000 7f93d620d900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-0/keyring: can't open /var/lib/ceph/osd/ceph-0/keyring: (2) No such file or directory 2026-04-01T13:54:11.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:11.889+0000 7f93d620d900 -1 created new key in keyring /var/lib/ceph/osd/ceph-0/keyring 2026-04-01T13:54:11.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:11.889+0000 7f93d620d900 -1 bdev(0x562bc9199800 /var/lib/ceph/osd/ceph-0/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:11.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:11.889+0000 7f93d620d900 -1 bluestore(/var/lib/ceph/osd/ceph-0) _read_fsid unparsable uuid 2026-04-01T13:54:12.569 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 2026-04-01T13:54:12.638 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 1 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:12.723 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:12.723+0000 7f2ce291b900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-1/keyring: can't open /var/lib/ceph/osd/ceph-1/keyring: (2) No such file or directory 2026-04-01T13:54:12.724 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:12.723+0000 7f2ce291b900 -1 created new key in keyring /var/lib/ceph/osd/ceph-1/keyring 2026-04-01T13:54:12.724 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:12.723+0000 7f2ce291b900 -1 bdev(0x56212703b800 /var/lib/ceph/osd/ceph-1/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:12.724 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:12.723+0000 7f2ce291b900 -1 bluestore(/var/lib/ceph/osd/ceph-1) _read_fsid unparsable uuid 2026-04-01T13:54:13.417 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 2026-04-01T13:54:13.441 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 2 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:13.521 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:13.521+0000 7f8b8a822900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-2/keyring: can't open /var/lib/ceph/osd/ceph-2/keyring: (2) No such file or directory 2026-04-01T13:54:13.521 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:13.521+0000 7f8b8a822900 -1 created new key in keyring /var/lib/ceph/osd/ceph-2/keyring 2026-04-01T13:54:13.522 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:13.521+0000 7f8b8a822900 -1 bdev(0x5626260b3800 /var/lib/ceph/osd/ceph-2/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:13.522 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:13.521+0000 7f8b8a822900 -1 bluestore(/var/lib/ceph/osd/ceph-2) _read_fsid unparsable uuid 2026-04-01T13:54:14.154 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 2026-04-01T13:54:14.222 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 3 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:14.308 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:14.307+0000 7f7db4494900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-3/keyring: can't open /var/lib/ceph/osd/ceph-3/keyring: (2) No such file or directory 2026-04-01T13:54:14.308 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:14.307+0000 7f7db4494900 -1 created new key in keyring /var/lib/ceph/osd/ceph-3/keyring 2026-04-01T13:54:14.308 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:14.307+0000 7f7db4494900 -1 bdev(0x564ebda55800 /var/lib/ceph/osd/ceph-3/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:14.308 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:14.307+0000 7f7db4494900 -1 bluestore(/var/lib/ceph/osd/ceph-3) _read_fsid unparsable uuid 2026-04-01T13:54:14.969 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 2026-04-01T13:54:14.994 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/ceph-4 2026-04-01T13:54:15.024 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:15.024 INFO:tasks.ceph:role: osd.4 2026-04-01T13:54:15.024 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm08.local 2026-04-01T13:54:15.024 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:15.094 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:15.099 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-04-01T13:54:15.102 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm08.local -o noatime 2026-04-01T13:54:15.102 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/ceph-4 2026-04-01T13:54:15.178 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-4 2026-04-01T13:54:15.251 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/ceph-5 2026-04-01T13:54:15.320 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:15.321 INFO:tasks.ceph:role: osd.5 2026-04-01T13:54:15.321 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm08.local 2026-04-01T13:54:15.321 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:15.387 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:15.393 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-04-01T13:54:15.395 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm08.local -o noatime 2026-04-01T13:54:15.395 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/ceph-5 2026-04-01T13:54:15.470 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-5 2026-04-01T13:54:15.544 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/ceph-6 2026-04-01T13:54:15.612 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:15.612 INFO:tasks.ceph:role: osd.6 2026-04-01T13:54:15.612 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm08.local 2026-04-01T13:54:15.612 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:15.676 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:15.681 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-04-01T13:54:15.683 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm08.local -o noatime 2026-04-01T13:54:15.683 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/ceph-6 2026-04-01T13:54:15.761 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-6 2026-04-01T13:54:15.829 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/osd/ceph-7 2026-04-01T13:54:15.896 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-01T13:54:15.896 INFO:tasks.ceph:role: osd.7 2026-04-01T13:54:15.896 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_4 on ubuntu@vm08.local 2026-04-01T13:54:15.896 DEBUG:teuthology.orchestra.run.vm08:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_4 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout:meta-data=/dev/vg_nvme/lv_4 isize=2048 agcount=4, agsize=1310464 blks 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout: = sunit=0 swidth=0 blks 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-01T13:54:15.962 INFO:teuthology.orchestra.run.vm08.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-01T13:54:15.967 INFO:teuthology.orchestra.run.vm08.stdout:Discarding blocks...Done. 2026-04-01T13:54:15.969 INFO:tasks.ceph:mount /dev/vg_nvme/lv_4 on ubuntu@vm08.local -o noatime 2026-04-01T13:54:15.969 DEBUG:teuthology.orchestra.run.vm08:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_4 /var/lib/ceph/osd/ceph-7 2026-04-01T13:54:16.042 DEBUG:teuthology.orchestra.run.vm08:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-7 2026-04-01T13:54:16.113 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 4 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:16.199 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.198+0000 7f9a5d8b3900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-4/keyring: can't open /var/lib/ceph/osd/ceph-4/keyring: (2) No such file or directory 2026-04-01T13:54:16.199 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.199+0000 7f9a5d8b3900 -1 created new key in keyring /var/lib/ceph/osd/ceph-4/keyring 2026-04-01T13:54:16.199 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.199+0000 7f9a5d8b3900 -1 bdev(0x55d4d52c3800 /var/lib/ceph/osd/ceph-4/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:16.199 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.199+0000 7f9a5d8b3900 -1 bluestore(/var/lib/ceph/osd/ceph-4) _read_fsid unparsable uuid 2026-04-01T13:54:16.852 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-4 2026-04-01T13:54:16.878 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 5 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:16.964 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.964+0000 7fea95ef7900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-5/keyring: can't open /var/lib/ceph/osd/ceph-5/keyring: (2) No such file or directory 2026-04-01T13:54:16.964 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.964+0000 7fea95ef7900 -1 created new key in keyring /var/lib/ceph/osd/ceph-5/keyring 2026-04-01T13:54:16.964 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.964+0000 7fea95ef7900 -1 bdev(0x56398b275800 /var/lib/ceph/osd/ceph-5/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:16.965 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:16.964+0000 7fea95ef7900 -1 bluestore(/var/lib/ceph/osd/ceph-5) _read_fsid unparsable uuid 2026-04-01T13:54:17.644 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 2026-04-01T13:54:17.670 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 6 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:17.753 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:17.752+0000 7f6e6200c900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-6/keyring: can't open /var/lib/ceph/osd/ceph-6/keyring: (2) No such file or directory 2026-04-01T13:54:17.753 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:17.752+0000 7f6e6200c900 -1 created new key in keyring /var/lib/ceph/osd/ceph-6/keyring 2026-04-01T13:54:17.753 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:17.752+0000 7f6e6200c900 -1 bdev(0x558ee4845800 /var/lib/ceph/osd/ceph-6/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:17.753 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:17.752+0000 7f6e6200c900 -1 bluestore(/var/lib/ceph/osd/ceph-6) _read_fsid unparsable uuid 2026-04-01T13:54:18.441 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 2026-04-01T13:54:18.508 DEBUG:teuthology.orchestra.run.vm08:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 7 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:18.588 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:18.588+0000 7f98abba0900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-7/keyring: can't open /var/lib/ceph/osd/ceph-7/keyring: (2) No such file or directory 2026-04-01T13:54:18.588 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:18.588+0000 7f98abba0900 -1 created new key in keyring /var/lib/ceph/osd/ceph-7/keyring 2026-04-01T13:54:18.588 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:18.588+0000 7f98abba0900 -1 bdev(0x560011835800 /var/lib/ceph/osd/ceph-7/block) open stat got: (1) Operation not permitted 2026-04-01T13:54:18.589 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:54:18.588+0000 7f98abba0900 -1 bluestore(/var/lib/ceph/osd/ceph-7) _read_fsid unparsable uuid 2026-04-01T13:54:19.189 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 2026-04-01T13:54:19.254 INFO:tasks.ceph:Reading keys from all nodes... 2026-04-01T13:54:19.264 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:19.264 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/mgr/ceph-y/keyring of=/dev/stdout 2026-04-01T13:54:19.294 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:19.294 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-0/keyring of=/dev/stdout 2026-04-01T13:54:19.362 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:19.362 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-1/keyring of=/dev/stdout 2026-04-01T13:54:19.428 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:19.429 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-2/keyring of=/dev/stdout 2026-04-01T13:54:19.493 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:19.493 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-3/keyring of=/dev/stdout 2026-04-01T13:54:19.559 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:19.559 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/mgr/ceph-x/keyring of=/dev/stdout 2026-04-01T13:54:19.588 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:19.588 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-4/keyring of=/dev/stdout 2026-04-01T13:54:19.654 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:19.654 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-5/keyring of=/dev/stdout 2026-04-01T13:54:19.720 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:19.720 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-6/keyring of=/dev/stdout 2026-04-01T13:54:19.785 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:19.786 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-7/keyring of=/dev/stdout 2026-04-01T13:54:19.851 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:19.851 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.client.0.keyring of=/dev/stdout 2026-04-01T13:54:19.871 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:19.871 DEBUG:teuthology.orchestra.run.vm08:> dd if=/etc/ceph/ceph.client.1.keyring of=/dev/stdout 2026-04-01T13:54:19.907 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-04-01T13:54:19.907 DEBUG:teuthology.orchestra.run.vm09:> dd if=/etc/ceph/ceph.client.2.keyring of=/dev/stdout 2026-04-01T13:54:19.924 INFO:tasks.ceph:Adding keys to all mons... 2026-04-01T13:54:19.924 DEBUG:teuthology.orchestra.run.vm06:> sudo tee -a /etc/ceph/ceph.keyring 2026-04-01T13:54:19.926 DEBUG:teuthology.orchestra.run.vm08:> sudo tee -a /etc/ceph/ceph.keyring 2026-04-01T13:54:19.950 INFO:teuthology.orchestra.run.vm06.stdout:[mgr.y] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCCI81p2Q7fIBAApnb872kvXhbGk4rb+rTH/A== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.0] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCDI81pl1kUNRAAhmapaEqq5gm0QgH0aUJrNQ== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.1] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCEI81pBqcxKxAA1saGghtorqsFvejoNxPTnw== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.2] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCFI81p//oiHxAALxWjvSo6aV9pyBbEPQKShw== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.3] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCGI81p10liEhAAup3BiQ+2JCrYakSIxsxT9g== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[mgr.x] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCCI81pGrnmIxAA0K+UMcXKoZobN7qcHjYc+g== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.4] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCII81p9BTpCxAA9ge74R+SA1xz15M9xylX9g== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.5] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCII81pqzuGORAAPDoqJCWF2OMM0+8Ocyi5vQ== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.6] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCJI81pm1npLBAANSpjynoJngjh3FiBdCXBdQ== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[osd.7] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCKI81pqlIeIxAAT7OVQsiG3h/s9FS92jvG8Q== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCCI81p2sVSJhAAzhzeUZhoF2I1vvwbxPsilg== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[client.1] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCCI81pax5CKRAA4WZAZ5eL7MulUUkkotXBKA== 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout:[client.2] 2026-04-01T13:54:19.956 INFO:teuthology.orchestra.run.vm06.stdout: key = AQCCI81pl2UGLBAAIqHc3DJptnPJq97sFNvQCQ== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[mgr.y] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCCI81p2Q7fIBAApnb872kvXhbGk4rb+rTH/A== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.0] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCDI81pl1kUNRAAhmapaEqq5gm0QgH0aUJrNQ== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.1] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCEI81pBqcxKxAA1saGghtorqsFvejoNxPTnw== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.2] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCFI81p//oiHxAALxWjvSo6aV9pyBbEPQKShw== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.3] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCGI81p10liEhAAup3BiQ+2JCrYakSIxsxT9g== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[mgr.x] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCCI81pGrnmIxAA0K+UMcXKoZobN7qcHjYc+g== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.4] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCII81p9BTpCxAA9ge74R+SA1xz15M9xylX9g== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.5] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCII81pqzuGORAAPDoqJCWF2OMM0+8Ocyi5vQ== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.6] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCJI81pm1npLBAANSpjynoJngjh3FiBdCXBdQ== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[osd.7] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCKI81pqlIeIxAAT7OVQsiG3h/s9FS92jvG8Q== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[client.0] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCCI81p2sVSJhAAzhzeUZhoF2I1vvwbxPsilg== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[client.1] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCCI81pax5CKRAA4WZAZ5eL7MulUUkkotXBKA== 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout:[client.2] 2026-04-01T13:54:19.972 INFO:teuthology.orchestra.run.vm08.stdout: key = AQCCI81pl2UGLBAAIqHc3DJptnPJq97sFNvQCQ== 2026-04-01T13:54:19.973 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.y --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-01T13:54:20.000 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.y --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-01T13:54:20.059 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.0 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.061 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.0 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.106 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.1 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.108 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.1 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.155 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.2 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.156 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.2 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.206 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.3 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.207 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.3 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.285 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.x --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-01T13:54:20.286 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.x --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-01T13:54:20.332 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.4 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.334 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.4 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.380 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.5 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.381 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.5 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.429 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.6 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.430 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.6 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.482 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.7 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.483 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.7 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-01T13:54:20.529 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.0 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-01T13:54:20.531 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.0 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-01T13:54:20.581 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.1 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-01T13:54:20.582 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.1 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-01T13:54:20.633 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.2 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-01T13:54:20.634 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.2 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-01T13:54:20.687 INFO:tasks.ceph:Running mkfs on mon nodes... 2026-04-01T13:54:20.687 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/mon/ceph-a 2026-04-01T13:54:20.712 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster ceph --mkfs -i a --monmap /home/ubuntu/cephtest/ceph.monmap --keyring /etc/ceph/ceph.keyring 2026-04-01T13:54:20.901 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/mon/ceph-a 2026-04-01T13:54:20.926 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/mon/ceph-c 2026-04-01T13:54:20.994 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster ceph --mkfs -i c --monmap /home/ubuntu/cephtest/ceph.monmap --keyring /etc/ceph/ceph.keyring 2026-04-01T13:54:21.159 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/mon/ceph-c 2026-04-01T13:54:21.186 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /var/lib/ceph/mon/ceph-b 2026-04-01T13:54:21.211 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster ceph --mkfs -i b --monmap /home/ubuntu/cephtest/ceph.monmap --keyring /etc/ceph/ceph.keyring 2026-04-01T13:54:21.493 DEBUG:teuthology.orchestra.run.vm08:> sudo chown -R ceph:ceph /var/lib/ceph/mon/ceph-b 2026-04-01T13:54:21.518 DEBUG:teuthology.orchestra.run.vm06:> rm -- /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:21.519 DEBUG:teuthology.orchestra.run.vm08:> rm -- /home/ubuntu/cephtest/ceph.monmap 2026-04-01T13:54:21.574 INFO:tasks.ceph:Starting mon daemons in cluster ceph... 2026-04-01T13:54:21.574 INFO:tasks.ceph.mon.a:Restarting daemon 2026-04-01T13:54:21.574 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i a 2026-04-01T13:54:21.577 INFO:tasks.ceph.mon.a:Started 2026-04-01T13:54:21.580 INFO:tasks.ceph.mon.c:Restarting daemon 2026-04-01T13:54:21.581 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i c 2026-04-01T13:54:21.582 INFO:tasks.ceph.mon.c:Started 2026-04-01T13:54:21.582 INFO:tasks.ceph.mon.b:Restarting daemon 2026-04-01T13:54:21.582 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i b 2026-04-01T13:54:21.615 INFO:tasks.ceph.mon.b:Started 2026-04-01T13:54:21.615 INFO:tasks.ceph:Starting mgr daemons in cluster ceph... 2026-04-01T13:54:21.616 INFO:tasks.ceph.mgr.y:Restarting daemon 2026-04-01T13:54:21.616 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster ceph -i y 2026-04-01T13:54:21.618 INFO:tasks.ceph.mgr.y:Started 2026-04-01T13:54:21.618 INFO:tasks.ceph.mgr.x:Restarting daemon 2026-04-01T13:54:21.618 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster ceph -i x 2026-04-01T13:54:21.619 INFO:tasks.ceph.mgr.x:Started 2026-04-01T13:54:21.619 DEBUG:tasks.ceph:set 0 configs 2026-04-01T13:54:21.619 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph config dump 2026-04-01T13:54:27.097 INFO:teuthology.orchestra.run.vm06.stdout:WHO MASK LEVEL OPTION VALUE RO 2026-04-01T13:54:27.109 INFO:tasks.ceph:Setting crush tunables to default 2026-04-01T13:54:27.109 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd crush tunables default 2026-04-01T13:54:27.248 INFO:teuthology.orchestra.run.vm06.stderr:adjusted tunables profile to default 2026-04-01T13:54:27.259 INFO:tasks.ceph:check_enable_crimson: False 2026-04-01T13:54:27.260 INFO:tasks.ceph:Starting osd daemons in cluster ceph... 2026-04-01T13:54:27.260 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:27.260 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-0/fsid of=/dev/stdout 2026-04-01T13:54:27.288 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:27.288 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-1/fsid of=/dev/stdout 2026-04-01T13:54:27.362 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:27.362 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-2/fsid of=/dev/stdout 2026-04-01T13:54:27.432 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:54:27.432 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-3/fsid of=/dev/stdout 2026-04-01T13:54:27.499 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:27.499 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-4/fsid of=/dev/stdout 2026-04-01T13:54:27.525 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:27.525 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-5/fsid of=/dev/stdout 2026-04-01T13:54:27.591 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:27.591 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-6/fsid of=/dev/stdout 2026-04-01T13:54:27.660 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-01T13:54:27.660 DEBUG:teuthology.orchestra.run.vm08:> sudo dd if=/var/lib/ceph/osd/ceph-7/fsid of=/dev/stdout 2026-04-01T13:54:27.729 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 55b1d571-4d43-479e-adcc-14d92adf6b2d 0 2026-04-01T13:54:27.787 INFO:tasks.ceph.mgr.x.vm08.stderr:/usr/lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-04-01T13:54:27.787 INFO:tasks.ceph.mgr.x.vm08.stderr:Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-04-01T13:54:27.787 INFO:tasks.ceph.mgr.x.vm08.stderr: from numpy import show_config as show_numpy_config 2026-04-01T13:54:27.892 INFO:teuthology.orchestra.run.vm08.stdout:0 2026-04-01T13:54:27.902 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 949773d5-b17f-4762-897e-481b4a3e8fd5 1 2026-04-01T13:54:28.067 INFO:teuthology.orchestra.run.vm08.stdout:1 2026-04-01T13:54:28.076 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 6ff87782-acbe-4022-bab6-2574d62f4f87 2 2026-04-01T13:54:28.193 INFO:teuthology.orchestra.run.vm08.stdout:2 2026-04-01T13:54:28.203 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 91f4b6f5-b474-4ab9-8d13-3aa7b4d4b428 3 2026-04-01T13:54:28.327 INFO:teuthology.orchestra.run.vm08.stdout:3 2026-04-01T13:54:28.342 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 5abade85-af9a-4b42-b355-6f718042997d 4 2026-04-01T13:54:28.449 INFO:tasks.ceph.mgr.y.vm06.stderr:/usr/lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-04-01T13:54:28.449 INFO:tasks.ceph.mgr.y.vm06.stderr:Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-04-01T13:54:28.449 INFO:tasks.ceph.mgr.y.vm06.stderr: from numpy import show_config as show_numpy_config 2026-04-01T13:54:28.464 INFO:teuthology.orchestra.run.vm08.stdout:4 2026-04-01T13:54:28.473 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 47637605-d9df-4b11-99fb-a4cb80a5c4cc 5 2026-04-01T13:54:28.591 INFO:teuthology.orchestra.run.vm08.stdout:5 2026-04-01T13:54:28.602 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 514e3574-9efd-496e-90e4-6299da5cdcdb 6 2026-04-01T13:54:28.729 INFO:teuthology.orchestra.run.vm08.stdout:6 2026-04-01T13:54:28.739 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph --cluster ceph osd new 09b9bbff-b98e-4938-98b1-b28ce091c042 7 2026-04-01T13:54:28.865 INFO:teuthology.orchestra.run.vm08.stdout:7 2026-04-01T13:54:28.875 INFO:tasks.ceph.osd.0:Restarting daemon 2026-04-01T13:54:28.875 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 0 2026-04-01T13:54:28.877 INFO:tasks.ceph.osd.0:Started 2026-04-01T13:54:28.877 INFO:tasks.ceph.osd.1:Restarting daemon 2026-04-01T13:54:28.877 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 1 2026-04-01T13:54:28.878 INFO:tasks.ceph.osd.1:Started 2026-04-01T13:54:28.878 INFO:tasks.ceph.osd.2:Restarting daemon 2026-04-01T13:54:28.878 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 2 2026-04-01T13:54:28.880 INFO:tasks.ceph.osd.2:Started 2026-04-01T13:54:28.880 INFO:tasks.ceph.osd.3:Restarting daemon 2026-04-01T13:54:28.880 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3 2026-04-01T13:54:28.881 INFO:tasks.ceph.osd.3:Started 2026-04-01T13:54:28.881 INFO:tasks.ceph.osd.4:Restarting daemon 2026-04-01T13:54:28.881 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 4 2026-04-01T13:54:28.883 INFO:tasks.ceph.osd.4:Started 2026-04-01T13:54:28.883 INFO:tasks.ceph.osd.5:Restarting daemon 2026-04-01T13:54:28.883 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 5 2026-04-01T13:54:28.885 INFO:tasks.ceph.osd.5:Started 2026-04-01T13:54:28.885 INFO:tasks.ceph.osd.6:Restarting daemon 2026-04-01T13:54:28.886 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 6 2026-04-01T13:54:28.889 INFO:tasks.ceph.osd.6:Started 2026-04-01T13:54:28.889 INFO:tasks.ceph.osd.7:Restarting daemon 2026-04-01T13:54:28.889 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 7 2026-04-01T13:54:28.893 INFO:tasks.ceph.osd.7:Started 2026-04-01T13:54:28.893 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json 2026-04-01T13:54:29.049 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T13:54:29.048+0000 7fc9e8bf3900 -1 Falling back to public interface 2026-04-01T13:54:29.050 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T13:54:29.049+0000 7f3313723900 -1 Falling back to public interface 2026-04-01T13:54:29.066 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T13:54:29.065+0000 7fee694ac900 -1 Falling back to public interface 2026-04-01T13:54:29.068 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T13:54:29.067+0000 7f8cba827900 -1 Falling back to public interface 2026-04-01T13:54:29.068 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:29.068 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":10,"fsid":"8a8efea3-ddf7-4150-909c-57db5b081e35","created":"2026-04-01T13:54:27.021551+0000","modified":"2026-04-01T13:54:28.861169+0000","last_up_change":"0.000000","last_in_change":"2026-04-01T13:54:28.861169+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"55b1d571-4d43-479e-adcc-14d92adf6b2d","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"949773d5-b17f-4762-897e-481b4a3e8fd5","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"6ff87782-acbe-4022-bab6-2574d62f4f87","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"91f4b6f5-b474-4ab9-8d13-3aa7b4d4b428","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"5abade85-af9a-4b42-b355-6f718042997d","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"47637605-d9df-4b11-99fb-a4cb80a5c4cc","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":6,"uuid":"514e3574-9efd-496e-90e4-6299da5cdcdb","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":7,"uuid":"09b9bbff-b98e-4938-98b1-b28ce091c042","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-01T13:54:29.072 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T13:54:29.071+0000 7fe85aeeb900 -1 Falling back to public interface 2026-04-01T13:54:29.079 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T13:54:29.079+0000 7fc64f2b9900 -1 Falling back to public interface 2026-04-01T13:54:29.080 INFO:tasks.ceph.ceph_manager.ceph:[] 2026-04-01T13:54:29.080 INFO:tasks.ceph:Waiting for OSDs to come up 2026-04-01T13:54:29.080 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T13:54:29.079+0000 7f7b6d57c900 -1 Falling back to public interface 2026-04-01T13:54:29.085 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T13:54:29.085+0000 7f68f3c80900 -1 Falling back to public interface 2026-04-01T13:54:29.508 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T13:54:29.507+0000 7fee694ac900 -1 osd.6 0 log_to_monitors true 2026-04-01T13:54:29.511 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T13:54:29.510+0000 7f3313723900 -1 osd.4 0 log_to_monitors true 2026-04-01T13:54:29.555 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T13:54:29.555+0000 7f68f3c80900 -1 osd.5 0 log_to_monitors true 2026-04-01T13:54:29.560 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T13:54:29.560+0000 7f8cba827900 -1 osd.3 0 log_to_monitors true 2026-04-01T13:54:29.567 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T13:54:29.567+0000 7fc9e8bf3900 -1 osd.2 0 log_to_monitors true 2026-04-01T13:54:29.568 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T13:54:29.568+0000 7fc64f2b9900 -1 osd.7 0 log_to_monitors true 2026-04-01T13:54:29.591 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T13:54:29.590+0000 7fe85aeeb900 -1 osd.1 0 log_to_monitors true 2026-04-01T13:54:29.631 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T13:54:29.631+0000 7f7b6d57c900 -1 osd.0 0 log_to_monitors true 2026-04-01T13:54:29.884 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster ceph osd dump --format=json 2026-04-01T13:54:29.995 INFO:teuthology.misc.health.vm06.stdout: 2026-04-01T13:54:29.995 INFO:teuthology.misc.health.vm06.stdout:{"epoch":10,"fsid":"8a8efea3-ddf7-4150-909c-57db5b081e35","created":"2026-04-01T13:54:27.021551+0000","modified":"2026-04-01T13:54:28.861169+0000","last_up_change":"0.000000","last_in_change":"2026-04-01T13:54:28.861169+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"55b1d571-4d43-479e-adcc-14d92adf6b2d","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"949773d5-b17f-4762-897e-481b4a3e8fd5","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"6ff87782-acbe-4022-bab6-2574d62f4f87","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"91f4b6f5-b474-4ab9-8d13-3aa7b4d4b428","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"5abade85-af9a-4b42-b355-6f718042997d","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"47637605-d9df-4b11-99fb-a4cb80a5c4cc","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":6,"uuid":"514e3574-9efd-496e-90e4-6299da5cdcdb","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":7,"uuid":"09b9bbff-b98e-4938-98b1-b28ce091c042","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-01T13:54:30.008 DEBUG:teuthology.misc:0 of 8 OSDs are up 2026-04-01T13:54:30.628 INFO:tasks.ceph.mgr.x.vm08.stderr:2026-04-01T13:54:30.628+0000 7f851888c640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-04-01T13:54:30.628 INFO:tasks.ceph.mgr.x.vm08.stderr:2026-04-01T13:54:30.628+0000 7f851888c640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-04-01T13:54:35.766 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T13:54:35.765+0000 7fc64b249640 -1 osd.7 0 waiting for initial osdmap 2026-04-01T13:54:35.781 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T13:54:35.781+0000 7fc64604d640 -1 osd.7 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.811 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T13:54:35.810+0000 7fee6543a640 -1 osd.6 0 waiting for initial osdmap 2026-04-01T13:54:35.817 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T13:54:35.816+0000 7f7b6950a640 -1 osd.0 0 waiting for initial osdmap 2026-04-01T13:54:35.824 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T13:54:35.824+0000 7fee6023e640 -1 osd.6 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.832 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T13:54:35.832+0000 7f7b6430e640 -1 osd.0 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.845 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T13:54:35.844+0000 7f8cb67e3640 -1 osd.3 0 waiting for initial osdmap 2026-04-01T13:54:35.860 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T13:54:35.859+0000 7fe856e79640 -1 osd.1 0 waiting for initial osdmap 2026-04-01T13:54:35.860 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T13:54:35.860+0000 7f8cb0dd5640 -1 osd.3 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.872 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T13:54:35.872+0000 7fe851c7d640 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.892 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T13:54:35.891+0000 7f330f6b1640 -1 osd.4 0 waiting for initial osdmap 2026-04-01T13:54:35.902 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T13:54:35.901+0000 7f330a4b5640 -1 osd.4 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.913 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T13:54:35.912+0000 7fc9e4b83640 -1 osd.2 0 waiting for initial osdmap 2026-04-01T13:54:35.927 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T13:54:35.927+0000 7fc9df987640 -1 osd.2 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:35.943 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T13:54:35.943+0000 7f68efc0e640 -1 osd.5 0 waiting for initial osdmap 2026-04-01T13:54:35.961 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T13:54:35.960+0000 7f68eaa12640 -1 osd.5 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-01T13:54:36.815 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster ceph osd dump --format=json 2026-04-01T13:54:37.009 INFO:teuthology.misc.health.vm06.stdout: 2026-04-01T13:54:37.009 INFO:teuthology.misc.health.vm06.stdout:{"epoch":13,"fsid":"8a8efea3-ddf7-4150-909c-57db5b081e35","created":"2026-04-01T13:54:27.021551+0000","modified":"2026-04-01T13:54:36.784319+0000","last_up_change":"2026-04-01T13:54:36.784319+0000","last_in_change":"2026-04-01T13:54:28.861169+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"55b1d571-4d43-479e-adcc-14d92adf6b2d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6825","nonce":4118484167}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6827","nonce":4118484167}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6831","nonce":4118484167}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6829","nonce":4118484167}]},"public_addr":"192.168.123.106:6825/4118484167","cluster_addr":"192.168.123.106:6827/4118484167","heartbeat_back_addr":"192.168.123.106:6831/4118484167","heartbeat_front_addr":"192.168.123.106:6829/4118484167","state":["exists","up"]},{"osd":1,"uuid":"949773d5-b17f-4762-897e-481b4a3e8fd5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6817","nonce":3442454584}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6819","nonce":3442454584}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6823","nonce":3442454584}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6821","nonce":3442454584}]},"public_addr":"192.168.123.106:6817/3442454584","cluster_addr":"192.168.123.106:6819/3442454584","heartbeat_back_addr":"192.168.123.106:6823/3442454584","heartbeat_front_addr":"192.168.123.106:6821/3442454584","state":["exists","up"]},{"osd":2,"uuid":"6ff87782-acbe-4022-bab6-2574d62f4f87","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6801","nonce":1320475448}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6803","nonce":1320475448}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6807","nonce":1320475448}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6805","nonce":1320475448}]},"public_addr":"192.168.123.106:6801/1320475448","cluster_addr":"192.168.123.106:6803/1320475448","heartbeat_back_addr":"192.168.123.106:6807/1320475448","heartbeat_front_addr":"192.168.123.106:6805/1320475448","state":["exists","up"]},{"osd":3,"uuid":"91f4b6f5-b474-4ab9-8d13-3aa7b4d4b428","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6809","nonce":3098443200}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6811","nonce":3098443200}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6815","nonce":3098443200}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6813","nonce":3098443200}]},"public_addr":"192.168.123.106:6809/3098443200","cluster_addr":"192.168.123.106:6811/3098443200","heartbeat_back_addr":"192.168.123.106:6815/3098443200","heartbeat_front_addr":"192.168.123.106:6813/3098443200","state":["exists","up"]},{"osd":4,"uuid":"5abade85-af9a-4b42-b355-6f718042997d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6801","nonce":4028317668}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6803","nonce":4028317668}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6807","nonce":4028317668}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6805","nonce":4028317668}]},"public_addr":"192.168.123.108:6801/4028317668","cluster_addr":"192.168.123.108:6803/4028317668","heartbeat_back_addr":"192.168.123.108:6807/4028317668","heartbeat_front_addr":"192.168.123.108:6805/4028317668","state":["exists","up"]},{"osd":5,"uuid":"47637605-d9df-4b11-99fb-a4cb80a5c4cc","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6825","nonce":3790433485}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6827","nonce":3790433485}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6831","nonce":3790433485}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6829","nonce":3790433485}]},"public_addr":"192.168.123.108:6825/3790433485","cluster_addr":"192.168.123.108:6827/3790433485","heartbeat_back_addr":"192.168.123.108:6831/3790433485","heartbeat_front_addr":"192.168.123.108:6829/3790433485","state":["exists","up"]},{"osd":6,"uuid":"514e3574-9efd-496e-90e4-6299da5cdcdb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6809","nonce":2183106347}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6811","nonce":2183106347}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6815","nonce":2183106347}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6813","nonce":2183106347}]},"public_addr":"192.168.123.108:6809/2183106347","cluster_addr":"192.168.123.108:6811/2183106347","heartbeat_back_addr":"192.168.123.108:6815/2183106347","heartbeat_front_addr":"192.168.123.108:6813/2183106347","state":["exists","up"]},{"osd":7,"uuid":"09b9bbff-b98e-4938-98b1-b28ce091c042","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6817","nonce":983645084}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6819","nonce":983645084}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6823","nonce":983645084}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6821","nonce":983645084}]},"public_addr":"192.168.123.108:6817/983645084","cluster_addr":"192.168.123.108:6819/983645084","heartbeat_back_addr":"192.168.123.108:6823/983645084","heartbeat_front_addr":"192.168.123.108:6821/983645084","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-01T13:54:37.020 DEBUG:teuthology.misc:8 of 8 OSDs are up 2026-04-01T13:54:37.020 INFO:tasks.ceph:Creating RBD pool 2026-04-01T13:54:37.020 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd pool create rbd 8 2026-04-01T13:54:38.127 INFO:teuthology.orchestra.run.vm06.stderr:pool 'rbd' created 2026-04-01T13:54:38.142 DEBUG:teuthology.orchestra.run.vm06:> rbd --cluster ceph pool init rbd 2026-04-01T13:54:38.176 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:54:38.176 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:54:41.149 INFO:tasks.ceph:Starting mds daemons in cluster ceph... 2026-04-01T13:54:41.150 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph config log 1 --format=json 2026-04-01T13:54:41.150 INFO:tasks.daemonwatchdog.daemon_watchdog:watchdog starting 2026-04-01T13:54:41.374 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:41.385 INFO:teuthology.orchestra.run.vm06.stdout:[{"version":9,"timestamp":"2026-04-01T13:54:35.944589+0000","name":"","changes":[{"name":"osd.5/osd_mclock_max_capacity_iops_hdd","new_value":"5111.363570"}]}] 2026-04-01T13:54:41.385 INFO:tasks.ceph_manager:config epoch is 9 2026-04-01T13:54:41.385 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-04-01T13:54:41.385 INFO:tasks.ceph.ceph_manager.ceph:waiting for mgr available 2026-04-01T13:54:41.385 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph mgr dump --format=json 2026-04-01T13:54:41.608 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:41.620 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":5,"flags":0,"active_gid":4102,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6832","nonce":3027633272},{"type":"v1","addr":"192.168.123.108:6833","nonce":3027633272}]},"active_addr":"192.168.123.108:6833/3027633272","active_change":"2026-04-01T13:54:29.610180+0000","active_mgr_features":4541880224203014143,"available":true,"standbys":[{"gid":4109,"name":"y","mgr_features":4541880224203014143,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.2.0","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":true,"error_string":"","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["iostat","nfs"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.2.0","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":true,"error_string":"","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"tentacle":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":0,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":1481245437}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":2319489127}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":1468507354}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.108:0","nonce":3035230872}]}]} 2026-04-01T13:54:41.621 INFO:tasks.ceph.ceph_manager.ceph:mgr available! 2026-04-01T13:54:41.621 INFO:tasks.ceph.ceph_manager.ceph:waiting for all up 2026-04-01T13:54:41.621 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json 2026-04-01T13:54:41.823 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:41.823 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":17,"fsid":"8a8efea3-ddf7-4150-909c-57db5b081e35","created":"2026-04-01T13:54:27.021551+0000","modified":"2026-04-01T13:54:41.135854+0000","last_up_change":"2026-04-01T13:54:36.784319+0000","last_in_change":"2026-04-01T13:54:28.861169+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":"rbd","create_time":"2026-04-01T13:54:37.242147+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"17","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":17,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2,"score_stable":2,"optimal_score":1,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":".mgr","create_time":"2026-04-01T13:54:37.639663+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"16","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":8,"score_stable":8,"optimal_score":0.25,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"55b1d571-4d43-479e-adcc-14d92adf6b2d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6825","nonce":4118484167}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6827","nonce":4118484167}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6831","nonce":4118484167}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6829","nonce":4118484167}]},"public_addr":"192.168.123.106:6825/4118484167","cluster_addr":"192.168.123.106:6827/4118484167","heartbeat_back_addr":"192.168.123.106:6831/4118484167","heartbeat_front_addr":"192.168.123.106:6829/4118484167","state":["exists","up"]},{"osd":1,"uuid":"949773d5-b17f-4762-897e-481b4a3e8fd5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6817","nonce":3442454584}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6819","nonce":3442454584}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6823","nonce":3442454584}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6821","nonce":3442454584}]},"public_addr":"192.168.123.106:6817/3442454584","cluster_addr":"192.168.123.106:6819/3442454584","heartbeat_back_addr":"192.168.123.106:6823/3442454584","heartbeat_front_addr":"192.168.123.106:6821/3442454584","state":["exists","up"]},{"osd":2,"uuid":"6ff87782-acbe-4022-bab6-2574d62f4f87","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6801","nonce":1320475448}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6803","nonce":1320475448}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6807","nonce":1320475448}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6805","nonce":1320475448}]},"public_addr":"192.168.123.106:6801/1320475448","cluster_addr":"192.168.123.106:6803/1320475448","heartbeat_back_addr":"192.168.123.106:6807/1320475448","heartbeat_front_addr":"192.168.123.106:6805/1320475448","state":["exists","up"]},{"osd":3,"uuid":"91f4b6f5-b474-4ab9-8d13-3aa7b4d4b428","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6809","nonce":3098443200}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6811","nonce":3098443200}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6815","nonce":3098443200}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6813","nonce":3098443200}]},"public_addr":"192.168.123.106:6809/3098443200","cluster_addr":"192.168.123.106:6811/3098443200","heartbeat_back_addr":"192.168.123.106:6815/3098443200","heartbeat_front_addr":"192.168.123.106:6813/3098443200","state":["exists","up"]},{"osd":4,"uuid":"5abade85-af9a-4b42-b355-6f718042997d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6801","nonce":4028317668}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6803","nonce":4028317668}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6807","nonce":4028317668}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6805","nonce":4028317668}]},"public_addr":"192.168.123.108:6801/4028317668","cluster_addr":"192.168.123.108:6803/4028317668","heartbeat_back_addr":"192.168.123.108:6807/4028317668","heartbeat_front_addr":"192.168.123.108:6805/4028317668","state":["exists","up"]},{"osd":5,"uuid":"47637605-d9df-4b11-99fb-a4cb80a5c4cc","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6825","nonce":3790433485}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6827","nonce":3790433485}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6831","nonce":3790433485}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6829","nonce":3790433485}]},"public_addr":"192.168.123.108:6825/3790433485","cluster_addr":"192.168.123.108:6827/3790433485","heartbeat_back_addr":"192.168.123.108:6831/3790433485","heartbeat_front_addr":"192.168.123.108:6829/3790433485","state":["exists","up"]},{"osd":6,"uuid":"514e3574-9efd-496e-90e4-6299da5cdcdb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6809","nonce":2183106347}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6811","nonce":2183106347}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6815","nonce":2183106347}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6813","nonce":2183106347}]},"public_addr":"192.168.123.108:6809/2183106347","cluster_addr":"192.168.123.108:6811/2183106347","heartbeat_back_addr":"192.168.123.108:6815/2183106347","heartbeat_front_addr":"192.168.123.108:6813/2183106347","state":["exists","up"]},{"osd":7,"uuid":"09b9bbff-b98e-4938-98b1-b28ce091c042","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6817","nonce":983645084}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6819","nonce":983645084}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6823","nonce":983645084}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6821","nonce":983645084}]},"public_addr":"192.168.123.108:6817/983645084","cluster_addr":"192.168.123.108:6819/983645084","heartbeat_back_addr":"192.168.123.108:6823/983645084","heartbeat_front_addr":"192.168.123.108:6821/983645084","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.640042+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.628742+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.586055+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.587924+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.464746+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.540595+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.468544+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.566587+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-01T13:54:41.834 INFO:tasks.ceph.ceph_manager.ceph:all up! 2026-04-01T13:54:41.834 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json 2026-04-01T13:54:42.032 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:42.032 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":17,"fsid":"8a8efea3-ddf7-4150-909c-57db5b081e35","created":"2026-04-01T13:54:27.021551+0000","modified":"2026-04-01T13:54:41.135854+0000","last_up_change":"2026-04-01T13:54:36.784319+0000","last_in_change":"2026-04-01T13:54:28.861169+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":"rbd","create_time":"2026-04-01T13:54:37.242147+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"17","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":17,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2,"score_stable":2,"optimal_score":1,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":".mgr","create_time":"2026-04-01T13:54:37.639663+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"16","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":8,"score_stable":8,"optimal_score":0.25,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"55b1d571-4d43-479e-adcc-14d92adf6b2d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6825","nonce":4118484167}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6827","nonce":4118484167}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6831","nonce":4118484167}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":4118484167},{"type":"v1","addr":"192.168.123.106:6829","nonce":4118484167}]},"public_addr":"192.168.123.106:6825/4118484167","cluster_addr":"192.168.123.106:6827/4118484167","heartbeat_back_addr":"192.168.123.106:6831/4118484167","heartbeat_front_addr":"192.168.123.106:6829/4118484167","state":["exists","up"]},{"osd":1,"uuid":"949773d5-b17f-4762-897e-481b4a3e8fd5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6817","nonce":3442454584}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6819","nonce":3442454584}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6823","nonce":3442454584}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":3442454584},{"type":"v1","addr":"192.168.123.106:6821","nonce":3442454584}]},"public_addr":"192.168.123.106:6817/3442454584","cluster_addr":"192.168.123.106:6819/3442454584","heartbeat_back_addr":"192.168.123.106:6823/3442454584","heartbeat_front_addr":"192.168.123.106:6821/3442454584","state":["exists","up"]},{"osd":2,"uuid":"6ff87782-acbe-4022-bab6-2574d62f4f87","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6801","nonce":1320475448}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6803","nonce":1320475448}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6807","nonce":1320475448}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":1320475448},{"type":"v1","addr":"192.168.123.106:6805","nonce":1320475448}]},"public_addr":"192.168.123.106:6801/1320475448","cluster_addr":"192.168.123.106:6803/1320475448","heartbeat_back_addr":"192.168.123.106:6807/1320475448","heartbeat_front_addr":"192.168.123.106:6805/1320475448","state":["exists","up"]},{"osd":3,"uuid":"91f4b6f5-b474-4ab9-8d13-3aa7b4d4b428","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6809","nonce":3098443200}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6811","nonce":3098443200}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6815","nonce":3098443200}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":3098443200},{"type":"v1","addr":"192.168.123.106:6813","nonce":3098443200}]},"public_addr":"192.168.123.106:6809/3098443200","cluster_addr":"192.168.123.106:6811/3098443200","heartbeat_back_addr":"192.168.123.106:6815/3098443200","heartbeat_front_addr":"192.168.123.106:6813/3098443200","state":["exists","up"]},{"osd":4,"uuid":"5abade85-af9a-4b42-b355-6f718042997d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6800","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6801","nonce":4028317668}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6802","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6803","nonce":4028317668}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6806","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6807","nonce":4028317668}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6804","nonce":4028317668},{"type":"v1","addr":"192.168.123.108:6805","nonce":4028317668}]},"public_addr":"192.168.123.108:6801/4028317668","cluster_addr":"192.168.123.108:6803/4028317668","heartbeat_back_addr":"192.168.123.108:6807/4028317668","heartbeat_front_addr":"192.168.123.108:6805/4028317668","state":["exists","up"]},{"osd":5,"uuid":"47637605-d9df-4b11-99fb-a4cb80a5c4cc","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6824","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6825","nonce":3790433485}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6826","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6827","nonce":3790433485}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6830","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6831","nonce":3790433485}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6828","nonce":3790433485},{"type":"v1","addr":"192.168.123.108:6829","nonce":3790433485}]},"public_addr":"192.168.123.108:6825/3790433485","cluster_addr":"192.168.123.108:6827/3790433485","heartbeat_back_addr":"192.168.123.108:6831/3790433485","heartbeat_front_addr":"192.168.123.108:6829/3790433485","state":["exists","up"]},{"osd":6,"uuid":"514e3574-9efd-496e-90e4-6299da5cdcdb","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6808","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6809","nonce":2183106347}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6810","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6811","nonce":2183106347}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6814","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6815","nonce":2183106347}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6812","nonce":2183106347},{"type":"v1","addr":"192.168.123.108:6813","nonce":2183106347}]},"public_addr":"192.168.123.108:6809/2183106347","cluster_addr":"192.168.123.108:6811/2183106347","heartbeat_back_addr":"192.168.123.108:6815/2183106347","heartbeat_front_addr":"192.168.123.108:6813/2183106347","state":["exists","up"]},{"osd":7,"uuid":"09b9bbff-b98e-4938-98b1-b28ce091c042","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":14,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6816","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6817","nonce":983645084}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6818","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6819","nonce":983645084}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6822","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6823","nonce":983645084}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.108:6820","nonce":983645084},{"type":"v1","addr":"192.168.123.108:6821","nonce":983645084}]},"public_addr":"192.168.123.108:6817/983645084","cluster_addr":"192.168.123.108:6819/983645084","heartbeat_back_addr":"192.168.123.108:6823/983645084","heartbeat_front_addr":"192.168.123.108:6821/983645084","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.640042+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.628742+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.586055+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.587924+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.464746+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.540595+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.468544+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-01T13:54:30.566587+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-01T13:54:42.043 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.0 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.1 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.2 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.3 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.4 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.5 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.6 flush_pg_stats 2026-04-01T13:54:42.044 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.7 flush_pg_stats 2026-04-01T13:54:42.236 INFO:teuthology.orchestra.run.vm06.stdout:55834574853 2026-04-01T13:54:42.236 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.3 2026-04-01T13:54:42.245 INFO:teuthology.orchestra.run.vm06.stdout:55834574853 2026-04-01T13:54:42.245 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.4 2026-04-01T13:54:42.254 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:42.255 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.7 2026-04-01T13:54:42.265 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:42.266 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.2 2026-04-01T13:54:42.268 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:42.268 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.1 2026-04-01T13:54:42.279 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:42.279 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.6 2026-04-01T13:54:42.284 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:42.284 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.0 2026-04-01T13:54:42.296 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:42.296 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.5 2026-04-01T13:54:42.583 INFO:teuthology.orchestra.run.vm06.stdout:55834574849 2026-04-01T13:54:42.601 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574849 for osd.2 2026-04-01T13:54:42.628 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:42.654 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574853 got 55834574851 for osd.3 2026-04-01T13:54:42.662 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:42.687 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.0 2026-04-01T13:54:42.687 INFO:teuthology.orchestra.run.vm06.stdout:55834574849 2026-04-01T13:54:42.689 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:42.703 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574849 for osd.6 2026-04-01T13:54:42.703 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:42.707 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:42.712 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:42.714 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.7 2026-04-01T13:54:42.721 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.1 2026-04-01T13:54:42.723 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574853 got 55834574851 for osd.4 2026-04-01T13:54:42.730 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.5 2026-04-01T13:54:43.601 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.2 2026-04-01T13:54:43.655 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.3 2026-04-01T13:54:43.688 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.0 2026-04-01T13:54:43.704 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.6 2026-04-01T13:54:43.714 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.7 2026-04-01T13:54:43.721 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.1 2026-04-01T13:54:43.723 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.4 2026-04-01T13:54:43.731 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.5 2026-04-01T13:54:43.868 INFO:teuthology.orchestra.run.vm06.stdout:55834574849 2026-04-01T13:54:43.903 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574849 for osd.2 2026-04-01T13:54:43.960 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:43.980 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574853 got 55834574851 for osd.3 2026-04-01T13:54:44.030 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:44.035 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:44.051 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.1 2026-04-01T13:54:44.051 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:44.051 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.0 2026-04-01T13:54:44.066 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.7 2026-04-01T13:54:44.075 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:44.087 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574853 got 55834574851 for osd.4 2026-04-01T13:54:44.119 INFO:teuthology.orchestra.run.vm06.stdout:55834574849 2026-04-01T13:54:44.133 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574849 for osd.6 2026-04-01T13:54:44.140 INFO:teuthology.orchestra.run.vm06.stdout:55834574850 2026-04-01T13:54:44.151 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574850 for osd.5 2026-04-01T13:54:44.903 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.2 2026-04-01T13:54:44.980 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.3 2026-04-01T13:54:45.051 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.1 2026-04-01T13:54:45.052 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.0 2026-04-01T13:54:45.067 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.7 2026-04-01T13:54:45.088 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.4 2026-04-01T13:54:45.107 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:45.133 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.6 2026-04-01T13:54:45.135 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.2 2026-04-01T13:54:45.135 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.151 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.5 2026-04-01T13:54:45.205 INFO:teuthology.orchestra.run.vm06.stdout:55834574853 2026-04-01T13:54:45.223 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574853 got 55834574853 for osd.3 2026-04-01T13:54:45.224 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.329 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:45.350 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.1 2026-04-01T13:54:45.350 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.382 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:45.384 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:45.387 INFO:teuthology.orchestra.run.vm06.stdout:55834574853 2026-04-01T13:54:45.398 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.0 2026-04-01T13:54:45.398 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.406 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.7 2026-04-01T13:54:45.406 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.411 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574853 got 55834574853 for osd.4 2026-04-01T13:54:45.411 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.443 INFO:teuthology.orchestra.run.vm06.stdout:55834574851 2026-04-01T13:54:45.455 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.6 2026-04-01T13:54:45.455 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.469 INFO:teuthology.orchestra.run.vm06.stdout:55834574852 2026-04-01T13:54:45.479 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.5 2026-04-01T13:54:45.479 DEBUG:teuthology.parallel:result is None 2026-04-01T13:54:45.479 INFO:tasks.ceph.ceph_manager.ceph:waiting for clean 2026-04-01T13:54:45.479 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json 2026-04-01T13:54:45.706 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:45.706 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-04-01T13:54:45.717 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":15,"stamp":"2026-04-01T13:54:45.623237+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":18,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":8,"kb":754974720,"kb_used":626640,"kb_used_data":2176,"kb_used_omap":50,"kb_used_meta":214477,"kb_avail":754348080,"statfs":{"total":773094113280,"available":772452433920,"internally_reserved":0,"allocated":2228224,"data_stored":1323654,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":51980,"internal_metadata":219624692},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"2.000233"},"pg_stats":[{"pgid":"1.7","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.494721+0000","last_change":"2026-04-01T13:54:41.494721+0000","last_active":"2026-04-01T13:54:41.494721+0000","last_peered":"2026-04-01T13:54:41.494721+0000","last_clean":"2026-04-01T13:54:41.494721+0000","last_became_active":"2026-04-01T13:54:39.134906+0000","last_became_peered":"2026-04-01T13:54:39.134906+0000","last_unstale":"2026-04-01T13:54:41.494721+0000","last_undegraded":"2026-04-01T13:54:41.494721+0000","last_fullsized":"2026-04-01T13:54:41.494721+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T14:28:56.023038+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00015501999999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,3],"acting":[6,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.6","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.494804+0000","last_change":"2026-04-01T13:54:41.494804+0000","last_active":"2026-04-01T13:54:41.494804+0000","last_peered":"2026-04-01T13:54:41.494804+0000","last_clean":"2026-04-01T13:54:41.494804+0000","last_became_active":"2026-04-01T13:54:39.134874+0000","last_became_peered":"2026-04-01T13:54:39.134874+0000","last_unstale":"2026-04-01T13:54:41.494804+0000","last_undegraded":"2026-04-01T13:54:41.494804+0000","last_fullsized":"2026-04-01T13:54:41.494804+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T15:10:50.886969+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00036408099999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0],"acting":[6,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.5","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.737343+0000","last_change":"2026-04-01T13:54:41.737343+0000","last_active":"2026-04-01T13:54:41.737343+0000","last_peered":"2026-04-01T13:54:41.737343+0000","last_clean":"2026-04-01T13:54:41.737343+0000","last_became_active":"2026-04-01T13:54:39.374371+0000","last_became_peered":"2026-04-01T13:54:39.374371+0000","last_unstale":"2026-04-01T13:54:41.737343+0000","last_undegraded":"2026-04-01T13:54:41.737343+0000","last_fullsized":"2026-04-01T13:54:41.737343+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00021763699999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2],"acting":[4,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.4","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.142960+0000","last_change":"2026-04-01T13:54:41.142960+0000","last_active":"2026-04-01T13:54:41.142960+0000","last_peered":"2026-04-01T13:54:41.142960+0000","last_clean":"2026-04-01T13:54:41.142960+0000","last_became_active":"2026-04-01T13:54:39.134600+0000","last_became_peered":"2026-04-01T13:54:39.134600+0000","last_unstale":"2026-04-01T13:54:41.142960+0000","last_undegraded":"2026-04-01T13:54:41.142960+0000","last_fullsized":"2026-04-01T13:54:41.142960+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00019034799999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5],"acting":[3,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"15'32","reported_seq":59,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.143489+0000","last_change":"2026-04-01T13:54:39.135579+0000","last_active":"2026-04-01T13:54:41.143489+0000","last_peered":"2026-04-01T13:54:41.143489+0000","last_clean":"2026-04-01T13:54:41.143489+0000","last_became_active":"2026-04-01T13:54:39.135402+0000","last_became_peered":"2026-04-01T13:54:39.135402+0000","last_unstale":"2026-04-01T13:54:41.143489+0000","last_undegraded":"2026-04-01T13:54:41.143489+0000","last_fullsized":"2026-04-01T13:54:41.143489+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T15:10:50.886969+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,1],"acting":[7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"1.3","version":"15'1","reported_seq":21,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.198671+0000","last_change":"2026-04-01T13:54:41.198671+0000","last_active":"2026-04-01T13:54:41.198671+0000","last_peered":"2026-04-01T13:54:41.198671+0000","last_clean":"2026-04-01T13:54:41.198671+0000","last_became_active":"2026-04-01T13:54:39.135117+0000","last_became_peered":"2026-04-01T13:54:39.135117+0000","last_unstale":"2026-04-01T13:54:41.198671+0000","last_undegraded":"2026-04-01T13:54:41.198671+0000","last_fullsized":"2026-04-01T13:54:41.198671+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00040009900000000001,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6],"acting":[1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.2","version":"17'2","reported_seq":22,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.146721+0000","last_change":"2026-04-01T13:54:41.143401+0000","last_active":"2026-04-01T13:54:41.146721+0000","last_peered":"2026-04-01T13:54:41.146721+0000","last_clean":"2026-04-01T13:54:41.146721+0000","last_became_active":"2026-04-01T13:54:39.136658+0000","last_became_peered":"2026-04-01T13:54:39.136658+0000","last_unstale":"2026-04-01T13:54:41.146721+0000","last_undegraded":"2026-04-01T13:54:41.146721+0000","last_fullsized":"2026-04-01T13:54:41.146721+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:12:30.000353+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00020823900000000001,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.1","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.474766+0000","last_change":"2026-04-01T13:54:41.474766+0000","last_active":"2026-04-01T13:54:41.474766+0000","last_peered":"2026-04-01T13:54:41.474766+0000","last_clean":"2026-04-01T13:54:41.474766+0000","last_became_active":"2026-04-01T13:54:39.137071+0000","last_became_peered":"2026-04-01T13:54:39.137071+0000","last_unstale":"2026-04-01T13:54:41.474766+0000","last_undegraded":"2026-04-01T13:54:41.474766+0000","last_fullsized":"2026-04-01T13:54:41.474766+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000211325,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,0],"acting":[5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.0","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.143584+0000","last_change":"2026-04-01T13:54:41.143584+0000","last_active":"2026-04-01T13:54:41.143584+0000","last_peered":"2026-04-01T13:54:41.143584+0000","last_clean":"2026-04-01T13:54:41.143584+0000","last_became_active":"2026-04-01T13:54:39.136629+0000","last_became_peered":"2026-04-01T13:54:39.136629+0000","last_unstale":"2026-04-01T13:54:41.143584+0000","last_undegraded":"2026-04-01T13:54:41.143584+0000","last_fullsized":"2026-04-01T13:54:41.143584+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T18:22:39.447383+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00028380000000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]}],"pool_stats":[{"poolid":2,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2},{"poolid":1,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":8}],"osd_stats":[{"osd":7,"up_from":13,"seq":55834574852,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27468,"kb_used_data":620,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344372,"statfs":{"total":96636764160,"available":96608636928,"internally_reserved":0,"allocated":634880,"data_stored":512839,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6822,"internal_metadata":27452762},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":13,"seq":55834574851,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27020,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344820,"statfs":{"total":96636764160,"available":96609095680,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6169,"internal_metadata":27453415},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":13,"seq":55834574852,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27020,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344820,"statfs":{"total":96636764160,"available":96609095680,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":13,"seq":55834574853,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":436620,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":93935220,"statfs":{"total":96636764160,"available":96189665280,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":13,"seq":55834574853,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27012,"kb_used_data":164,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344828,"statfs":{"total":96636764160,"available":96609103872,"internally_reserved":0,"allocated":167936,"data_stored":53540,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6825,"internal_metadata":27452759},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":13,"seq":55834574851,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27020,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344820,"statfs":{"total":96636764160,"available":96609095680,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574852,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27464,"kb_used_data":616,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344376,"statfs":{"total":96636764160,"available":96608641024,"internally_reserved":0,"allocated":630784,"data_stored":512820,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6827,"internal_metadata":27452757},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":13,"seq":55834574852,"num_pgs":4,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27016,"kb_used_data":168,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344824,"statfs":{"total":96636764160,"available":96609099776,"internally_reserved":0,"allocated":172032,"data_stored":53559,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6827,"internal_metadata":27452757},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-04-01T13:54:45.717 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json 2026-04-01T13:54:45.901 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:45.901 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-04-01T13:54:45.913 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":15,"stamp":"2026-04-01T13:54:45.623237+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":18,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":8,"kb":754974720,"kb_used":626640,"kb_used_data":2176,"kb_used_omap":50,"kb_used_meta":214477,"kb_avail":754348080,"statfs":{"total":773094113280,"available":772452433920,"internally_reserved":0,"allocated":2228224,"data_stored":1323654,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":51980,"internal_metadata":219624692},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"2.000233"},"pg_stats":[{"pgid":"1.7","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.494721+0000","last_change":"2026-04-01T13:54:41.494721+0000","last_active":"2026-04-01T13:54:41.494721+0000","last_peered":"2026-04-01T13:54:41.494721+0000","last_clean":"2026-04-01T13:54:41.494721+0000","last_became_active":"2026-04-01T13:54:39.134906+0000","last_became_peered":"2026-04-01T13:54:39.134906+0000","last_unstale":"2026-04-01T13:54:41.494721+0000","last_undegraded":"2026-04-01T13:54:41.494721+0000","last_fullsized":"2026-04-01T13:54:41.494721+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T14:28:56.023038+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00015501999999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,3],"acting":[6,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.6","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.494804+0000","last_change":"2026-04-01T13:54:41.494804+0000","last_active":"2026-04-01T13:54:41.494804+0000","last_peered":"2026-04-01T13:54:41.494804+0000","last_clean":"2026-04-01T13:54:41.494804+0000","last_became_active":"2026-04-01T13:54:39.134874+0000","last_became_peered":"2026-04-01T13:54:39.134874+0000","last_unstale":"2026-04-01T13:54:41.494804+0000","last_undegraded":"2026-04-01T13:54:41.494804+0000","last_fullsized":"2026-04-01T13:54:41.494804+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T15:10:50.886969+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00036408099999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0],"acting":[6,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.5","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.737343+0000","last_change":"2026-04-01T13:54:41.737343+0000","last_active":"2026-04-01T13:54:41.737343+0000","last_peered":"2026-04-01T13:54:41.737343+0000","last_clean":"2026-04-01T13:54:41.737343+0000","last_became_active":"2026-04-01T13:54:39.374371+0000","last_became_peered":"2026-04-01T13:54:39.374371+0000","last_unstale":"2026-04-01T13:54:41.737343+0000","last_undegraded":"2026-04-01T13:54:41.737343+0000","last_fullsized":"2026-04-01T13:54:41.737343+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00021763699999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2],"acting":[4,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.4","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.142960+0000","last_change":"2026-04-01T13:54:41.142960+0000","last_active":"2026-04-01T13:54:41.142960+0000","last_peered":"2026-04-01T13:54:41.142960+0000","last_clean":"2026-04-01T13:54:41.142960+0000","last_became_active":"2026-04-01T13:54:39.134600+0000","last_became_peered":"2026-04-01T13:54:39.134600+0000","last_unstale":"2026-04-01T13:54:41.142960+0000","last_undegraded":"2026-04-01T13:54:41.142960+0000","last_fullsized":"2026-04-01T13:54:41.142960+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00019034799999999999,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5],"acting":[3,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"15'32","reported_seq":59,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.143489+0000","last_change":"2026-04-01T13:54:39.135579+0000","last_active":"2026-04-01T13:54:41.143489+0000","last_peered":"2026-04-01T13:54:41.143489+0000","last_clean":"2026-04-01T13:54:41.143489+0000","last_became_active":"2026-04-01T13:54:39.135402+0000","last_became_peered":"2026-04-01T13:54:39.135402+0000","last_unstale":"2026-04-01T13:54:41.143489+0000","last_undegraded":"2026-04-01T13:54:41.143489+0000","last_fullsized":"2026-04-01T13:54:41.143489+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T15:10:50.886969+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,1],"acting":[7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"1.3","version":"15'1","reported_seq":21,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.198671+0000","last_change":"2026-04-01T13:54:41.198671+0000","last_active":"2026-04-01T13:54:41.198671+0000","last_peered":"2026-04-01T13:54:41.198671+0000","last_clean":"2026-04-01T13:54:41.198671+0000","last_became_active":"2026-04-01T13:54:39.135117+0000","last_became_peered":"2026-04-01T13:54:39.135117+0000","last_unstale":"2026-04-01T13:54:41.198671+0000","last_undegraded":"2026-04-01T13:54:41.198671+0000","last_fullsized":"2026-04-01T13:54:41.198671+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00040009900000000001,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6],"acting":[1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.2","version":"17'2","reported_seq":22,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.146721+0000","last_change":"2026-04-01T13:54:41.143401+0000","last_active":"2026-04-01T13:54:41.146721+0000","last_peered":"2026-04-01T13:54:41.146721+0000","last_clean":"2026-04-01T13:54:41.146721+0000","last_became_active":"2026-04-01T13:54:39.136658+0000","last_became_peered":"2026-04-01T13:54:39.136658+0000","last_unstale":"2026-04-01T13:54:41.146721+0000","last_undegraded":"2026-04-01T13:54:41.146721+0000","last_fullsized":"2026-04-01T13:54:41.146721+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:12:30.000353+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00020823900000000001,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.1","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.474766+0000","last_change":"2026-04-01T13:54:41.474766+0000","last_active":"2026-04-01T13:54:41.474766+0000","last_peered":"2026-04-01T13:54:41.474766+0000","last_clean":"2026-04-01T13:54:41.474766+0000","last_became_active":"2026-04-01T13:54:39.137071+0000","last_became_peered":"2026-04-01T13:54:39.137071+0000","last_unstale":"2026-04-01T13:54:41.474766+0000","last_undegraded":"2026-04-01T13:54:41.474766+0000","last_fullsized":"2026-04-01T13:54:41.474766+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T20:04:19.584385+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000211325,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,0],"acting":[5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.0","version":"0'0","reported_seq":20,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-01T13:54:41.143584+0000","last_change":"2026-04-01T13:54:41.143584+0000","last_active":"2026-04-01T13:54:41.143584+0000","last_peered":"2026-04-01T13:54:41.143584+0000","last_clean":"2026-04-01T13:54:41.143584+0000","last_became_active":"2026-04-01T13:54:39.136629+0000","last_became_peered":"2026-04-01T13:54:39.136629+0000","last_unstale":"2026-04-01T13:54:41.143584+0000","last_undegraded":"2026-04-01T13:54:41.143584+0000","last_fullsized":"2026-04-01T13:54:41.143584+0000","mapping_epoch":14,"log_start":"0'0","ondisk_log_start":"0'0","created":14,"last_epoch_clean":15,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-01T13:54:38.117885+0000","last_clean_scrub_stamp":"2026-04-01T13:54:38.117885+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-02T18:22:39.447383+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00028380000000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]}],"pool_stats":[{"poolid":2,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2},{"poolid":1,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":8}],"osd_stats":[{"osd":7,"up_from":13,"seq":55834574852,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27468,"kb_used_data":620,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344372,"statfs":{"total":96636764160,"available":96608636928,"internally_reserved":0,"allocated":634880,"data_stored":512839,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6822,"internal_metadata":27452762},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":13,"seq":55834574851,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27020,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344820,"statfs":{"total":96636764160,"available":96609095680,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6169,"internal_metadata":27453415},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":13,"seq":55834574852,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27020,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344820,"statfs":{"total":96636764160,"available":96609095680,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":13,"seq":55834574853,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":436620,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":93935220,"statfs":{"total":96636764160,"available":96189665280,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":13,"seq":55834574853,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27012,"kb_used_data":164,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344828,"statfs":{"total":96636764160,"available":96609103872,"internally_reserved":0,"allocated":167936,"data_stored":53540,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6825,"internal_metadata":27452759},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":13,"seq":55834574851,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27020,"kb_used_data":152,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344820,"statfs":{"total":96636764160,"available":96609095680,"internally_reserved":0,"allocated":155648,"data_stored":47724,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574852,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27464,"kb_used_data":616,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344376,"statfs":{"total":96636764160,"available":96608641024,"internally_reserved":0,"allocated":630784,"data_stored":512820,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6827,"internal_metadata":27452757},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":13,"seq":55834574852,"num_pgs":4,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27016,"kb_used_data":168,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344824,"statfs":{"total":96636764160,"available":96609099776,"internally_reserved":0,"allocated":172032,"data_stored":53559,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6827,"internal_metadata":27452757},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-04-01T13:54:45.913 INFO:tasks.ceph.ceph_manager.ceph:clean! 2026-04-01T13:54:45.913 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-04-01T13:54:45.913 INFO:tasks.ceph.ceph_manager.ceph:wait_until_healthy 2026-04-01T13:54:45.913 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph health --format=json 2026-04-01T13:54:46.126 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:54:46.126 INFO:teuthology.orchestra.run.vm06.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-04-01T13:54:46.138 INFO:tasks.ceph.ceph_manager.ceph:wait_until_healthy done 2026-04-01T13:54:46.138 INFO:teuthology.run_tasks:Running task openssl_keys... 2026-04-01T13:54:46.141 INFO:teuthology.run_tasks:Running task rgw... 2026-04-01T13:54:46.144 DEBUG:tasks.rgw:config is {'client.0': None, 'client.1': None, 'client.2': None} 2026-04-01T13:54:46.144 DEBUG:tasks.rgw:client list is dict_keys(['client.0', 'client.1', 'client.2']) 2026-04-01T13:54:46.145 INFO:tasks.rgw:Creating data pools 2026-04-01T13:54:46.145 DEBUG:tasks.rgw:Obtaining remote for client client.0 2026-04-01T13:54:46.145 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster ceph 2026-04-01T13:54:46.712 INFO:teuthology.orchestra.run.vm06.stderr:pool 'default.rgw.buckets.data' created 2026-04-01T13:54:46.735 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster ceph 2026-04-01T13:54:47.741 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-04-01T13:54:47.779 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster ceph 2026-04-01T13:54:48.747 INFO:teuthology.orchestra.run.vm06.stderr:pool 'default.rgw.buckets.index' created 2026-04-01T13:54:48.768 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster ceph 2026-04-01T13:54:49.747 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-04-01T13:54:49.784 DEBUG:tasks.rgw:Obtaining remote for client client.1 2026-04-01T13:54:49.784 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster ceph 2026-04-01T13:54:49.983 INFO:teuthology.orchestra.run.vm08.stderr:pool 'default.rgw.buckets.data' already exists 2026-04-01T13:54:49.995 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster ceph 2026-04-01T13:54:50.752 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-04-01T13:54:50.765 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster ceph 2026-04-01T13:54:50.951 INFO:teuthology.orchestra.run.vm08.stderr:pool 'default.rgw.buckets.index' already exists 2026-04-01T13:54:50.962 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster ceph 2026-04-01T13:54:51.781 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-04-01T13:54:51.793 DEBUG:tasks.rgw:Obtaining remote for client client.2 2026-04-01T13:54:51.793 DEBUG:teuthology.orchestra.run.vm09:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster ceph 2026-04-01T13:54:51.984 INFO:teuthology.orchestra.run.vm09.stderr:pool 'default.rgw.buckets.data' already exists 2026-04-01T13:54:51.996 DEBUG:teuthology.orchestra.run.vm09:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster ceph 2026-04-01T13:54:52.787 INFO:teuthology.orchestra.run.vm09.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-04-01T13:54:52.800 DEBUG:teuthology.orchestra.run.vm09:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster ceph 2026-04-01T13:54:52.986 INFO:teuthology.orchestra.run.vm09.stderr:pool 'default.rgw.buckets.index' already exists 2026-04-01T13:54:52.997 DEBUG:teuthology.orchestra.run.vm09:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster ceph 2026-04-01T13:54:53.802 INFO:teuthology.orchestra.run.vm09.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-04-01T13:54:53.815 DEBUG:tasks.rgw:Pools created 2026-04-01T13:54:53.815 INFO:tasks.util.rgw:rgwadmin: client.0 : ['user', 'list'] 2026-04-01T13:54:53.815 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'user', 'list'] 2026-04-01T13:54:53.815 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph user list 2026-04-01T13:54:53.850 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:54:53.850 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:54:55.855 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.854+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 realm 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:55.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.856+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.857 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.857+0000 7f8edff52900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:54:55.857 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.857+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.858 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.857+0000 7f8edff52900 20 rados_obj.operate() r=0 bl.length=1060 2026-04-01T13:54:55.858 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.857+0000 7f8edff52900 20 searching for the correct realm 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 RGWRados::pool_iterate: got zone_info.28f6e1de-cca5-4d4f-a2d4-c22f622f50fe 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 RGWRados::pool_iterate: got zonegroup_info.6d3bd279-f200-47e0-be3a-2ea7a6959538 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 RGWRados::pool_iterate: got default.zone. 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.866+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 20 rados_obj.operate() r=0 bl.length=436 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 20 zone default found 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 4 Realm: () 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 4 ZoneGroup: default (6d3bd279-f200-47e0-be3a-2ea7a6959538) 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 4 Zone: default (28f6e1de-cca5-4d4f-a2d4-c22f622f50fe) 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 20 zonegroup default 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:55.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:55.867+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:57.848 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:57.847+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:57.848 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:57.847+0000 7f8edff52900 20 rados->read ofs=0 len=0 2026-04-01T13:54:57.849 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:57.849+0000 7f8edff52900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:54:57.849 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:57.849+0000 7f8edff52900 20 started sync module instance, tier type = 2026-04-01T13:54:57.849 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:57.849+0000 7f8edff52900 20 started zone id=28f6e1de-cca5-4d4f-a2d4-c22f622f50fe (name=default) with tier type = 2026-04-01T13:54:59.856 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.855+0000 7f8edff52900 20 add_watcher() i=3 2026-04-01T13:54:59.857 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.856+0000 7f8edff52900 20 add_watcher() i=0 2026-04-01T13:54:59.858 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.858+0000 7f8edff52900 20 add_watcher() i=6 2026-04-01T13:54:59.858 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.858+0000 7f8edff52900 20 add_watcher() i=2 2026-04-01T13:54:59.859 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.858+0000 7f8edff52900 20 add_watcher() i=5 2026-04-01T13:54:59.861 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.861+0000 7f8edff52900 20 add_watcher() i=4 2026-04-01T13:54:59.861 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.861+0000 7f8edff52900 20 add_watcher() i=1 2026-04-01T13:54:59.862 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.862+0000 7f8edff52900 20 add_watcher() i=7 2026-04-01T13:54:59.862 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.862+0000 7f8edff52900 2 all 8 watchers are set, enabling cache 2026-04-01T13:54:59.863 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.863+0000 7f8ed4ff9640 5 boost::asio::awaitable, obj_version> > logback_generations::read(const DoutPrefixProvider*):446: oid=data_loggenerations_metadata not found 2026-04-01T13:54:59.863 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.863+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.0 2026-04-01T13:54:59.863 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.863+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.864 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.864+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.0 does not exist 2026-04-01T13:54:59.864 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.864+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.1 2026-04-01T13:54:59.864 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.864+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.865 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.864+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.1 does not exist 2026-04-01T13:54:59.865 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.864+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.2 2026-04-01T13:54:59.865 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.864+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.865 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.865+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.2 does not exist 2026-04-01T13:54:59.865 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.865+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.3 2026-04-01T13:54:59.865 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.865+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.865+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.3 does not exist 2026-04-01T13:54:59.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.865+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.4 2026-04-01T13:54:59.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.865+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.4 does not exist 2026-04-01T13:54:59.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.5 2026-04-01T13:54:59.866 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.5 does not exist 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.6 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.6 does not exist 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.7 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.866+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.7 does not exist 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.8 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.8 does not exist 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.9 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.9 does not exist 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.10 2026-04-01T13:54:59.867 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.867+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.10 does not exist 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.11 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.11 does not exist 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.12 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.12 does not exist 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.13 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.13 does not exist 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.14 2026-04-01T13:54:59.868 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.14 does not exist 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.15 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.868+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.15 does not exist 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.16 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.16 does not exist 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.17 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.17 does not exist 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.18 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.18 does not exist 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.19 2026-04-01T13:54:59.869 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.19 does not exist 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.20 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.869+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.20 does not exist 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.21 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.21 does not exist 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.22 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.22 does not exist 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.23 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.23 does not exist 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.24 2026-04-01T13:54:59.870 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.24 does not exist 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.25 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.870+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.25 does not exist 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.26 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.26 does not exist 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.27 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.27 does not exist 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.28 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.28 does not exist 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.29 2026-04-01T13:54:59.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.29 does not exist 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.30 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.871+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.30 does not exist 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.31 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.31 does not exist 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.32 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.32 does not exist 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.33 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.33 does not exist 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.34 2026-04-01T13:54:59.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.34 does not exist 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.35 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.872+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.35 does not exist 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.36 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.36 does not exist 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.37 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.37 does not exist 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.38 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.38 does not exist 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.39 2026-04-01T13:54:59.873 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.39 does not exist 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.40 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.873+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.40 does not exist 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.41 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.41 does not exist 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.42 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.42 does not exist 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.43 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.43 does not exist 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.44 2026-04-01T13:54:59.874 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.44 does not exist 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.45 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.874+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.45 does not exist 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.46 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.46 does not exist 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.47 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.47 does not exist 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.48 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.48 does not exist 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.49 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.49 does not exist 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.50 2026-04-01T13:54:59.875 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.875+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.50 does not exist 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.51 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.51 does not exist 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.52 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.52 does not exist 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.53 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.53 does not exist 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.54 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.54 does not exist 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.55 2026-04-01T13:54:59.876 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.55 does not exist 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.56 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.876+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.56 does not exist 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.57 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.57 does not exist 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.58 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.58 does not exist 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.59 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.59 does not exist 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.60 2026-04-01T13:54:59.877 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.60 does not exist 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.61 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.877+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.61 does not exist 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.62 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.62 does not exist 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.63 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.63 does not exist 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.64 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.64 does not exist 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.65 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.65 does not exist 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.66 2026-04-01T13:54:59.878 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.66 does not exist 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.67 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.878+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.67 does not exist 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.68 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.68 does not exist 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.69 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.69 does not exist 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.70 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.70 does not exist 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.71 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.879 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.71 does not exist 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.72 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.72 does not exist 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.73 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.879+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.73 does not exist 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.74 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.74 does not exist 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.75 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.75 does not exist 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.76 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.76 does not exist 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.77 2026-04-01T13:54:59.880 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.77 does not exist 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.78 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.880+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.78 does not exist 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.79 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.79 does not exist 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.80 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.80 does not exist 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.81 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.81 does not exist 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.82 2026-04-01T13:54:59.881 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.82 does not exist 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.83 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.881+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.83 does not exist 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.84 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.84 does not exist 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.85 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.85 does not exist 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.86 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.86 does not exist 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.87 2026-04-01T13:54:59.882 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.87 does not exist 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.88 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.88 does not exist 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.89 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.882+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.89 does not exist 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.90 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.90 does not exist 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.91 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.91 does not exist 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.92 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.92 does not exist 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.93 2026-04-01T13:54:59.883 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.93 does not exist 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.94 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.883+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.94 does not exist 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.95 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.95 does not exist 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.96 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.96 does not exist 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.97 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.97 does not exist 2026-04-01T13:54:59.884 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.98 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.98 does not exist 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.99 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.884+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.99 does not exist 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.100 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.100 does not exist 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.101 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.101 does not exist 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.102 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.885 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.102 does not exist 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.103 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.103 does not exist 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.104 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.885+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.104 does not exist 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.105 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.105 does not exist 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.106 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.106 does not exist 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.107 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.107 does not exist 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.108 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.886 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.108 does not exist 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.109 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.886+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.109 does not exist 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.110 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.110 does not exist 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.111 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.111 does not exist 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.112 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.112 does not exist 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.113 2026-04-01T13:54:59.887 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.113 does not exist 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.114 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.887+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.114 does not exist 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.115 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.115 does not exist 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.116 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.116 does not exist 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.117 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.117 does not exist 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.118 2026-04-01T13:54:59.888 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.118 does not exist 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.119 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.888+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.119 does not exist 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.120 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ed4ff9640 20 do_open: entering 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.120 does not exist 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ecffff640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.121 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.121 does not exist 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ecf7fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.122 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ecf7fe640 20 do_open: entering 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.122 does not exist 2026-04-01T13:54:59.889 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8edcd67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.123 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8edcd67640 20 do_open: entering 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.123 does not exist 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ed67fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.124 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.889+0000 7f8ed67fc640 20 do_open: entering 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.124 does not exist 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed5ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.125 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed5ffb640 20 do_open: entering 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.125 does not exist 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed57fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.126 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed57fa640 20 do_open: entering 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.126 does not exist 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8eddff0640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.127 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8eddff0640 20 do_open: entering 2026-04-01T13:54:59.890 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed4ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.127 does not exist 2026-04-01T13:54:59.891 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.890+0000 7f8ed4ff9640 20 do_create: entering 2026-04-01T13:54:59.892 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.892+0000 7f8ecffff640 20 do_open: entering 2026-04-01T13:54:59.895 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.894+0000 7f8edff52900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:54:59.895 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:54:59.894+0000 7f8edff52900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:02.896 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.896+0000 7f8edff52900 10 rgw_init_ioctx warning: failed to set recovery_priority on default.rgw.meta 2026-04-01T13:55:02.896 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.896+0000 7f8edff52900 5 note: GC not initialized 2026-04-01T13:55:02.896 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.896+0000 7f8e84ff1640 20 reqs_thread_entry: start 2026-04-01T13:55:02.958 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.957+0000 7f8edff52900 20 init_complete bucket index max shards: 11 2026-04-01T13:55:02.958 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.957+0000 7f8edff52900 20 Filter name: none 2026-04-01T13:55:02.958 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.957+0000 7f8e7e7fc640 20 reqs_thread_entry: start 2026-04-01T13:55:02.967 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.967+0000 7f8edff52900 20 remove_watcher() i=2 2026-04-01T13:55:02.967 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.967+0000 7f8edff52900 2 removed watcher, disabling cache 2026-04-01T13:55:02.967 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.967+0000 7f8edff52900 20 remove_watcher() i=0 2026-04-01T13:55:02.968 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.968+0000 7f8edff52900 20 remove_watcher() i=3 2026-04-01T13:55:02.968 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.968+0000 7f8edff52900 20 remove_watcher() i=1 2026-04-01T13:55:02.968 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.968+0000 7f8edff52900 20 remove_watcher() i=6 2026-04-01T13:55:02.968 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.968+0000 7f8edff52900 20 remove_watcher() i=5 2026-04-01T13:55:02.969 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.968+0000 7f8edff52900 20 remove_watcher() i=7 2026-04-01T13:55:02.969 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:02.968+0000 7f8edff52900 20 remove_watcher() i=4 2026-04-01T13:55:02.975 INFO:teuthology.orchestra.run.vm06.stdout:[] 2026-04-01T13:55:02.975 DEBUG:tasks.util.rgw: json result: [] 2026-04-01T13:55:02.975 INFO:tasks.rgw:Configuring storage class = FROZEN 2026-04-01T13:55:02.975 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-01T13:55:02.975 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-01T13:55:02.975 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN 2026-04-01T13:55:03.055 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.055 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.070 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.070+0000 7f91f2212900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.070 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.070+0000 7f91f2212900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.070 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.070+0000 7f919b7ee640 20 reqs_thread_entry: start 2026-04-01T13:55:03.080 INFO:teuthology.orchestra.run.vm06.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","STANDARD"]}}] 2026-04-01T13:55:03.080 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'STANDARD']}}] 2026-04-01T13:55:03.080 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-01T13:55:03.080 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-01T13:55:03.080 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN --data-pool default.rgw.buckets.data.frozen 2026-04-01T13:55:03.165 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.166 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.181 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.180+0000 7fa7bc80f900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.181 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.180+0000 7fa7bc80f900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.181 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.180+0000 7fa764ff1640 20 reqs_thread_entry: start 2026-04-01T13:55:03.193 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"28f6e1de-cca5-4d4f-a2d4-c22f622f50fe","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-01T13:55:03.193 DEBUG:tasks.util.rgw: json result: {'id': '28f6e1de-cca5-4d4f-a2d4-c22f622f50fe', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-01T13:55:03.193 INFO:tasks.rgw:Configuring storage class = LUKEWARM 2026-04-01T13:55:03.193 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-01T13:55:03.193 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-01T13:55:03.193 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM 2026-04-01T13:55:03.231 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.231 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.247 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.246+0000 7f6962f2c900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.247 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.246+0000 7f6962f2c900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.247 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.246+0000 7f6907fff640 20 reqs_thread_entry: start 2026-04-01T13:55:03.256 INFO:teuthology.orchestra.run.vm06.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-01T13:55:03.256 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-01T13:55:03.256 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-01T13:55:03.257 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-01T13:55:03.257 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM --data-pool default.rgw.buckets.data.lukewarm 2026-04-01T13:55:03.335 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.335 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.347 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.347+0000 7faea8b34900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.347 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.347+0000 7faea8b34900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.347 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:03.347+0000 7fae527ec640 20 reqs_thread_entry: start 2026-04-01T13:55:03.356 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"28f6e1de-cca5-4d4f-a2d4-c22f622f50fe","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-01T13:55:03.356 DEBUG:tasks.util.rgw: json result: {'id': '28f6e1de-cca5-4d4f-a2d4-c22f622f50fe', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-01T13:55:03.356 INFO:tasks.util.rgw:rgwadmin: client.1 : ['user', 'list'] 2026-04-01T13:55:03.356 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'user', 'list'] 2026-04-01T13:55:03.357 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph user list 2026-04-01T13:55:03.393 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.393 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.410 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.410+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.411+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.411+0000 7f6d6d968900 20 realm 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.411+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:55:03.412 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.412+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.413 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.413+0000 7f6d6d968900 20 rados_obj.operate() r=0 bl.length=1190 2026-04-01T13:55:03.413 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.413+0000 7f6d6d968900 20 searching for the correct realm 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 RGWRados::pool_iterate: got zone_info.28f6e1de-cca5-4d4f-a2d4-c22f622f50fe 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 RGWRados::pool_iterate: got zonegroup_info.6d3bd279-f200-47e0-be3a-2ea7a6959538 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 RGWRados::pool_iterate: got default.zone. 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:55:03.421 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados_obj.operate() r=0 bl.length=470 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 zone default found 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 4 Realm: () 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 4 ZoneGroup: default (6d3bd279-f200-47e0-be3a-2ea7a6959538) 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 4 Zone: default (28f6e1de-cca5-4d4f-a2d4-c22f622f50fe) 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 zonegroup default 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.421+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 rados->read ofs=0 len=0 2026-04-01T13:55:03.422 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:03.423 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 started sync module instance, tier type = 2026-04-01T13:55:03.423 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.422+0000 7f6d6d968900 20 started zone id=28f6e1de-cca5-4d4f-a2d4-c22f622f50fe (name=default) with tier type = 2026-04-01T13:55:03.425 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.425+0000 7f6d6d968900 20 add_watcher() i=0 2026-04-01T13:55:03.425 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.425+0000 7f6d6d968900 20 add_watcher() i=4 2026-04-01T13:55:03.426 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.425+0000 7f6d6d968900 20 add_watcher() i=1 2026-04-01T13:55:03.426 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.425+0000 7f6d6d968900 20 add_watcher() i=2 2026-04-01T13:55:03.426 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.426+0000 7f6d6d968900 20 add_watcher() i=3 2026-04-01T13:55:03.426 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.426+0000 7f6d6d968900 20 add_watcher() i=5 2026-04-01T13:55:03.426 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.426+0000 7f6d6d968900 20 add_watcher() i=6 2026-04-01T13:55:03.427 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.427+0000 7f6d6d968900 20 add_watcher() i=7 2026-04-01T13:55:03.427 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.427+0000 7f6d6d968900 2 all 8 watchers are set, enabling cache 2026-04-01T13:55:03.429 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.428+0000 7f6d6d968900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.429 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.428+0000 7f6d6d968900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.429 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.428+0000 7f6d6d968900 5 note: GC not initialized 2026-04-01T13:55:03.429 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.429+0000 7f6d167ec640 20 reqs_thread_entry: start 2026-04-01T13:55:03.567 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.566+0000 7f6d6d968900 20 init_complete bucket index max shards: 11 2026-04-01T13:55:03.567 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.566+0000 7f6d6d968900 20 Filter name: none 2026-04-01T13:55:03.567 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.566+0000 7f6d0ffff640 20 reqs_thread_entry: start 2026-04-01T13:55:03.575 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.574+0000 7f6d6d968900 20 remove_watcher() i=0 2026-04-01T13:55:03.575 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.574+0000 7f6d6d968900 2 removed watcher, disabling cache 2026-04-01T13:55:03.575 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.575+0000 7f6d6d968900 20 remove_watcher() i=1 2026-04-01T13:55:03.575 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.575+0000 7f6d6d968900 20 remove_watcher() i=5 2026-04-01T13:55:03.575 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.575+0000 7f6d6d968900 20 remove_watcher() i=3 2026-04-01T13:55:03.575 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.575+0000 7f6d6d968900 20 remove_watcher() i=6 2026-04-01T13:55:03.576 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.575+0000 7f6d6d968900 20 remove_watcher() i=2 2026-04-01T13:55:03.576 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.575+0000 7f6d6d968900 20 remove_watcher() i=4 2026-04-01T13:55:03.576 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.576+0000 7f6d6d968900 20 remove_watcher() i=7 2026-04-01T13:55:03.581 INFO:teuthology.orchestra.run.vm08.stdout:[] 2026-04-01T13:55:03.581 DEBUG:tasks.util.rgw: json result: [] 2026-04-01T13:55:03.581 INFO:tasks.rgw:Configuring storage class = FROZEN 2026-04-01T13:55:03.582 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-01T13:55:03.582 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-01T13:55:03.582 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN 2026-04-01T13:55:03.659 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.659 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.673 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.672+0000 7f5f6db3e900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.673 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.672+0000 7f5f6db3e900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.673 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.673+0000 7f5f177ee640 20 reqs_thread_entry: start 2026-04-01T13:55:03.682 INFO:teuthology.orchestra.run.vm08.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-01T13:55:03.683 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-01T13:55:03.683 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-01T13:55:03.683 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-01T13:55:03.683 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN --data-pool default.rgw.buckets.data.frozen 2026-04-01T13:55:03.770 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.770 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.783 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.782+0000 7f67db992900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.783 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.782+0000 7f67db992900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.783 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.783+0000 7f6784fe9640 20 reqs_thread_entry: start 2026-04-01T13:55:03.793 INFO:teuthology.orchestra.run.vm08.stdout:{"id":"28f6e1de-cca5-4d4f-a2d4-c22f622f50fe","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-01T13:55:03.793 DEBUG:tasks.util.rgw: json result: {'id': '28f6e1de-cca5-4d4f-a2d4-c22f622f50fe', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-01T13:55:03.793 INFO:tasks.rgw:Configuring storage class = LUKEWARM 2026-04-01T13:55:03.793 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-01T13:55:03.793 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-01T13:55:03.793 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM 2026-04-01T13:55:03.833 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.833 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.846 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.846+0000 7fa89db54900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.846 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.846+0000 7fa89db54900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.847 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.846+0000 7fa8477ee640 20 reqs_thread_entry: start 2026-04-01T13:55:03.855 INFO:teuthology.orchestra.run.vm08.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-01T13:55:03.855 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-01T13:55:03.855 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-01T13:55:03.855 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-01T13:55:03.855 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM --data-pool default.rgw.buckets.data.lukewarm 2026-04-01T13:55:03.934 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.934 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:03.951 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.950+0000 7f3dca122900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:03.951 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.950+0000 7f3dca122900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:03.951 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-01T13:55:03.951+0000 7f3d737ee640 20 reqs_thread_entry: start 2026-04-01T13:55:03.961 INFO:teuthology.orchestra.run.vm08.stdout:{"id":"28f6e1de-cca5-4d4f-a2d4-c22f622f50fe","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-01T13:55:03.961 DEBUG:tasks.util.rgw: json result: {'id': '28f6e1de-cca5-4d4f-a2d4-c22f622f50fe', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-01T13:55:03.961 INFO:tasks.util.rgw:rgwadmin: client.2 : ['user', 'list'] 2026-04-01T13:55:03.961 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'user', 'list'] 2026-04-01T13:55:03.961 DEBUG:teuthology.orchestra.run.vm09:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph user list 2026-04-01T13:55:03.996 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:03.996 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:04.013 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.013+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.014+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.014+0000 7fd51b53e900 20 realm 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.014+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.014+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.014+0000 7fd51b53e900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.014+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.015+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.014 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.015+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.015 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.015+0000 7fd51b53e900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:55:04.015 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.015+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.016 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.016+0000 7fd51b53e900 20 rados_obj.operate() r=0 bl.length=1190 2026-04-01T13:55:04.016 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.016+0000 7fd51b53e900 20 searching for the correct realm 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 RGWRados::pool_iterate: got zone_info.28f6e1de-cca5-4d4f-a2d4-c22f622f50fe 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 RGWRados::pool_iterate: got zonegroup_info.6d3bd279-f200-47e0-be3a-2ea7a6959538 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 RGWRados::pool_iterate: got default.zone. 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-01T13:55:04.024 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados_obj.operate() r=0 bl.length=470 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 zone default found 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 4 Realm: () 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 4 ZoneGroup: default (6d3bd279-f200-47e0-be3a-2ea7a6959538) 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 4 Zone: default (28f6e1de-cca5-4d4f-a2d4-c22f622f50fe) 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 zonegroup default 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.025+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.025 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 rados->read ofs=0 len=0 2026-04-01T13:55:04.026 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:04.026 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 started sync module instance, tier type = 2026-04-01T13:55:04.026 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.026+0000 7fd51b53e900 20 started zone id=28f6e1de-cca5-4d4f-a2d4-c22f622f50fe (name=default) with tier type = 2026-04-01T13:55:04.029 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.029+0000 7fd51b53e900 20 add_watcher() i=3 2026-04-01T13:55:04.029 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.029+0000 7fd51b53e900 20 add_watcher() i=0 2026-04-01T13:55:04.029 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.029+0000 7fd51b53e900 20 add_watcher() i=7 2026-04-01T13:55:04.029 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.030+0000 7fd51b53e900 20 add_watcher() i=5 2026-04-01T13:55:04.029 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.030+0000 7fd51b53e900 20 add_watcher() i=6 2026-04-01T13:55:04.030 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.030+0000 7fd51b53e900 20 add_watcher() i=4 2026-04-01T13:55:04.030 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.031+0000 7fd51b53e900 20 add_watcher() i=1 2026-04-01T13:55:04.030 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.031+0000 7fd51b53e900 20 add_watcher() i=2 2026-04-01T13:55:04.030 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.031+0000 7fd51b53e900 2 all 8 watchers are set, enabling cache 2026-04-01T13:55:04.032 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.032+0000 7fd51b53e900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:04.032 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.032+0000 7fd51b53e900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:04.032 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.032+0000 7fd51b53e900 5 note: GC not initialized 2026-04-01T13:55:04.033 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.033+0000 7fd4c4fe9640 20 reqs_thread_entry: start 2026-04-01T13:55:04.078 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.078+0000 7fd51b53e900 20 init_complete bucket index max shards: 11 2026-04-01T13:55:04.078 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.078+0000 7fd51b53e900 20 Filter name: none 2026-04-01T13:55:04.078 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.078+0000 7fd4be7fc640 20 reqs_thread_entry: start 2026-04-01T13:55:04.088 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.089+0000 7fd51b53e900 20 remove_watcher() i=7 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.089+0000 7fd51b53e900 2 removed watcher, disabling cache 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.089+0000 7fd51b53e900 20 remove_watcher() i=6 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.089+0000 7fd51b53e900 20 remove_watcher() i=0 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.089+0000 7fd51b53e900 20 remove_watcher() i=2 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.089+0000 7fd51b53e900 20 remove_watcher() i=5 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.090+0000 7fd51b53e900 20 remove_watcher() i=3 2026-04-01T13:55:04.089 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.090+0000 7fd51b53e900 20 remove_watcher() i=4 2026-04-01T13:55:04.090 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.090+0000 7fd51b53e900 20 remove_watcher() i=1 2026-04-01T13:55:04.096 INFO:teuthology.orchestra.run.vm09.stdout:[] 2026-04-01T13:55:04.096 DEBUG:tasks.util.rgw: json result: [] 2026-04-01T13:55:04.096 INFO:tasks.rgw:Configuring storage class = FROZEN 2026-04-01T13:55:04.096 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-01T13:55:04.096 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-01T13:55:04.096 DEBUG:teuthology.orchestra.run.vm09:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN 2026-04-01T13:55:04.177 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:04.177 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:04.192 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.192+0000 7fb05ec26900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:04.192 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.192+0000 7fb05ec26900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:04.192 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.192+0000 7fb0067ec640 20 reqs_thread_entry: start 2026-04-01T13:55:04.202 INFO:teuthology.orchestra.run.vm09.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-01T13:55:04.202 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-01T13:55:04.202 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-01T13:55:04.202 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-01T13:55:04.202 DEBUG:teuthology.orchestra.run.vm09:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN --data-pool default.rgw.buckets.data.frozen 2026-04-01T13:55:04.282 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:04.282 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:04.295 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.295+0000 7f9ca993a900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:04.295 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.295+0000 7f9ca993a900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:04.295 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.296+0000 7f9c52fed640 20 reqs_thread_entry: start 2026-04-01T13:55:04.308 INFO:teuthology.orchestra.run.vm09.stdout:{"id":"28f6e1de-cca5-4d4f-a2d4-c22f622f50fe","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-01T13:55:04.308 DEBUG:tasks.util.rgw: json result: {'id': '28f6e1de-cca5-4d4f-a2d4-c22f622f50fe', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-01T13:55:04.308 INFO:tasks.rgw:Configuring storage class = LUKEWARM 2026-04-01T13:55:04.308 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-01T13:55:04.308 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-01T13:55:04.308 DEBUG:teuthology.orchestra.run.vm09:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM 2026-04-01T13:55:04.389 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:04.389 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:04.404 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.404+0000 7f6b77752900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:04.404 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.404+0000 7f6b77752900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:04.404 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.404+0000 7f6b22fed640 20 reqs_thread_entry: start 2026-04-01T13:55:04.412 INFO:teuthology.orchestra.run.vm09.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-01T13:55:04.413 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-01T13:55:04.413 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-01T13:55:04.413 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-01T13:55:04.413 DEBUG:teuthology.orchestra.run.vm09:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM --data-pool default.rgw.buckets.data.lukewarm 2026-04-01T13:55:04.493 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:04.494 INFO:teuthology.orchestra.run.vm09.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:04.509 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.510+0000 7f2e8153e900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:04.509 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.510+0000 7f2e8153e900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:04.510 INFO:teuthology.orchestra.run.vm09.stderr:2026-04-01T13:55:04.510+0000 7f2e2afed640 20 reqs_thread_entry: start 2026-04-01T13:55:04.522 INFO:teuthology.orchestra.run.vm09.stdout:{"id":"28f6e1de-cca5-4d4f-a2d4-c22f622f50fe","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-01T13:55:04.522 DEBUG:tasks.util.rgw: json result: {'id': '28f6e1de-cca5-4d4f-a2d4-c22f622f50fe', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-01T13:55:04.522 INFO:tasks.rgw:Starting rgw... 2026-04-01T13:55:04.522 INFO:tasks.rgw:rgw client.0 config is {} 2026-04-01T13:55:04.522 INFO:tasks.rgw:Using beast as radosgw frontend 2026-04-01T13:55:04.522 DEBUG:teuthology.orchestra.run.vm06:> sudo echo -n http://vm06.local:80 | sudo tee /home/ubuntu/cephtest/url_file 2026-04-01T13:55:04.548 INFO:teuthology.orchestra.run.vm06.stdout:http://vm06.local:80 2026-04-01T13:55:04.548 DEBUG:teuthology.orchestra.run.vm06:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-04-01T13:55:04.611 INFO:tasks.rgw.client.0:Restarting daemon 2026-04-01T13:55:04.611 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.0 --cluster ceph -k /etc/ceph/ceph.client.0.keyring --log-file /var/log/ceph/rgw.ceph.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.0.stdout 2>&1 2026-04-01T13:55:04.653 INFO:tasks.rgw.client.0:Started 2026-04-01T13:55:04.653 INFO:tasks.rgw:rgw client.1 config is {} 2026-04-01T13:55:04.653 INFO:tasks.rgw:Using beast as radosgw frontend 2026-04-01T13:55:04.653 DEBUG:teuthology.orchestra.run.vm08:> sudo echo -n http://vm08.local:80 | sudo tee /home/ubuntu/cephtest/url_file 2026-04-01T13:55:04.681 INFO:teuthology.orchestra.run.vm08.stdout:http://vm08.local:80 2026-04-01T13:55:04.681 DEBUG:teuthology.orchestra.run.vm08:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-04-01T13:55:04.747 INFO:tasks.rgw.client.1:Restarting daemon 2026-04-01T13:55:04.748 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.1 --cluster ceph -k /etc/ceph/ceph.client.1.keyring --log-file /var/log/ceph/rgw.ceph.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.1.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.1.stdout 2>&1 2026-04-01T13:55:04.789 INFO:tasks.rgw.client.1:Started 2026-04-01T13:55:04.789 INFO:tasks.rgw:rgw client.2 config is {} 2026-04-01T13:55:04.789 INFO:tasks.rgw:Using beast as radosgw frontend 2026-04-01T13:55:04.789 DEBUG:teuthology.orchestra.run.vm09:> sudo echo -n http://vm09.local:80 | sudo tee /home/ubuntu/cephtest/url_file 2026-04-01T13:55:04.818 INFO:teuthology.orchestra.run.vm09.stdout:http://vm09.local:80 2026-04-01T13:55:04.818 DEBUG:teuthology.orchestra.run.vm09:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-04-01T13:55:04.890 INFO:tasks.rgw.client.2:Restarting daemon 2026-04-01T13:55:04.890 DEBUG:teuthology.orchestra.run.vm09:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.2 --cluster ceph -k /etc/ceph/ceph.client.2.keyring --log-file /var/log/ceph/rgw.ceph.client.2.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.2.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.2.stdout 2>&1 2026-04-01T13:55:04.932 INFO:tasks.rgw.client.2:Started 2026-04-01T13:55:04.932 INFO:tasks.rgw:Polling client.0 until it starts accepting connections on http://vm06.local:80/ 2026-04-01T13:55:04.932 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:80/ 2026-04-01T13:55:04.972 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-01T13:55:04.972 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-04-01T13:55:04.974 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 93500 0 --:--:-- --:--:-- --:--:-- 62333 2026-04-01T13:55:04.975 INFO:teuthology.orchestra.run.vm06.stdout:anonymous 2026-04-01T13:55:04.975 INFO:tasks.rgw:Polling client.1 until it starts accepting connections on http://vm08.local:80/ 2026-04-01T13:55:04.975 DEBUG:teuthology.orchestra.run.vm08:> curl http://vm08.local:80/ 2026-04-01T13:55:05.007 INFO:teuthology.orchestra.run.vm08.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-01T13:55:05.007 INFO:teuthology.orchestra.run.vm08.stderr: Dload Upload Total Spent Left Speed 2026-04-01T13:55:05.008 INFO:teuthology.orchestra.run.vm08.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-04-01T13:55:05.008 INFO:teuthology.orchestra.run.vm08.stderr:curl: (7) Failed to connect to vm08.local port 80: Connection refused 2026-04-01T13:55:05.009 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-04-01T13:55:06.011 DEBUG:teuthology.orchestra.run.vm08:> curl http://vm08.local:80/ 2026-04-01T13:55:06.032 INFO:teuthology.orchestra.run.vm08.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-01T13:55:06.032 INFO:teuthology.orchestra.run.vm08.stderr: Dload Upload Total Spent Left Speed 2026-04-01T13:55:06.033 INFO:teuthology.orchestra.run.vm08.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 93500 0 --:--:-- --:--:-- --:--:-- 93500 2026-04-01T13:55:06.034 INFO:teuthology.orchestra.run.vm08.stdout:anonymous 2026-04-01T13:55:06.034 INFO:tasks.rgw:Polling client.2 until it starts accepting connections on http://vm09.local:80/ 2026-04-01T13:55:06.034 DEBUG:teuthology.orchestra.run.vm09:> curl http://vm09.local:80/ 2026-04-01T13:55:06.058 INFO:teuthology.orchestra.run.vm09.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-01T13:55:06.058 INFO:teuthology.orchestra.run.vm09.stderr: Dload Upload Total Spent Left Speed 2026-04-01T13:55:06.059 INFO:teuthology.orchestra.run.vm09.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-04-01T13:55:06.060 INFO:teuthology.orchestra.run.vm09.stdout:anonymous 2026-04-01T13:55:06.060 INFO:teuthology.run_tasks:Running task tox... 2026-04-01T13:55:06.062 INFO:tasks.tox:Deploying tox from pip... 2026-04-01T13:55:06.062 DEBUG:teuthology.orchestra.run.vm06:> python3 -m venv /home/ubuntu/cephtest/tox-venv 2026-04-01T13:55:07.338 DEBUG:teuthology.orchestra.run.vm06:> source /home/ubuntu/cephtest/tox-venv/bin/activate && pip install tox 2026-04-01T13:55:07.644 INFO:teuthology.orchestra.run.vm06.stdout:Collecting tox 2026-04-01T13:55:07.673 INFO:teuthology.orchestra.run.vm06.stdout: Downloading tox-4.30.3-py3-none-any.whl (175 kB) 2026-04-01T13:55:07.748 INFO:teuthology.orchestra.run.vm06.stdout:Collecting chardet>=5.2 2026-04-01T13:55:07.757 INFO:teuthology.orchestra.run.vm06.stdout: Downloading chardet-5.2.0-py3-none-any.whl (199 kB) 2026-04-01T13:55:07.785 INFO:teuthology.orchestra.run.vm06.stdout:Collecting pyproject-api>=1.9.1 2026-04-01T13:55:07.793 INFO:teuthology.orchestra.run.vm06.stdout: Downloading pyproject_api-1.9.1-py3-none-any.whl (13 kB) 2026-04-01T13:55:07.885 INFO:teuthology.orchestra.run.vm06.stdout:Collecting virtualenv>=20.31.2 2026-04-01T13:55:07.893 INFO:teuthology.orchestra.run.vm06.stdout: Downloading virtualenv-21.2.0-py3-none-any.whl (5.8 MB) 2026-04-01T13:55:07.997 INFO:teuthology.orchestra.run.vm06.stdout:Collecting cachetools>=6.1 2026-04-01T13:55:08.005 INFO:teuthology.orchestra.run.vm06.stdout: Downloading cachetools-6.2.6-py3-none-any.whl (11 kB) 2026-04-01T13:55:08.030 INFO:teuthology.orchestra.run.vm06.stdout:Collecting colorama>=0.4.6 2026-04-01T13:55:08.038 INFO:teuthology.orchestra.run.vm06.stdout: Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB) 2026-04-01T13:55:08.068 INFO:teuthology.orchestra.run.vm06.stdout:Collecting packaging>=25 2026-04-01T13:55:08.077 INFO:teuthology.orchestra.run.vm06.stdout: Downloading packaging-26.0-py3-none-any.whl (74 kB) 2026-04-01T13:55:08.112 INFO:teuthology.orchestra.run.vm06.stdout:Collecting platformdirs>=4.3.8 2026-04-01T13:55:08.121 INFO:teuthology.orchestra.run.vm06.stdout: Downloading platformdirs-4.4.0-py3-none-any.whl (18 kB) 2026-04-01T13:55:08.158 INFO:teuthology.orchestra.run.vm06.stdout:Collecting filelock>=3.18 2026-04-01T13:55:08.167 INFO:teuthology.orchestra.run.vm06.stdout: Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2026-04-01T13:55:08.198 INFO:teuthology.orchestra.run.vm06.stdout:Collecting typing-extensions>=4.14.1 2026-04-01T13:55:08.206 INFO:teuthology.orchestra.run.vm06.stdout: Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2026-04-01T13:55:08.230 INFO:teuthology.orchestra.run.vm06.stdout:Collecting pluggy>=1.6 2026-04-01T13:55:08.238 INFO:teuthology.orchestra.run.vm06.stdout: Downloading pluggy-1.6.0-py3-none-any.whl (20 kB) 2026-04-01T13:55:08.291 INFO:teuthology.orchestra.run.vm06.stdout:Collecting tomli>=2.2.1 2026-04-01T13:55:08.299 INFO:teuthology.orchestra.run.vm06.stdout: Downloading tomli-2.4.1-py3-none-any.whl (14 kB) 2026-04-01T13:55:08.336 INFO:teuthology.orchestra.run.vm06.stdout:Collecting python-discovery>=1 2026-04-01T13:55:08.345 INFO:teuthology.orchestra.run.vm06.stdout: Downloading python_discovery-1.2.1-py3-none-any.whl (31 kB) 2026-04-01T13:55:08.371 INFO:teuthology.orchestra.run.vm06.stdout:Collecting distlib<1,>=0.3.7 2026-04-01T13:55:08.381 INFO:teuthology.orchestra.run.vm06.stdout: Downloading distlib-0.4.0-py2.py3-none-any.whl (469 kB) 2026-04-01T13:55:08.443 INFO:teuthology.orchestra.run.vm06.stdout:Installing collected packages: platformdirs, filelock, typing-extensions, tomli, python-discovery, packaging, distlib, virtualenv, pyproject-api, pluggy, colorama, chardet, cachetools, tox 2026-04-01T13:55:09.121 INFO:teuthology.orchestra.run.vm06.stdout:Successfully installed cachetools-6.2.6 chardet-5.2.0 colorama-0.4.6 distlib-0.4.0 filelock-3.19.1 packaging-26.0 platformdirs-4.4.0 pluggy-1.6.0 pyproject-api-1.9.1 python-discovery-1.2.1 tomli-2.4.1 tox-4.30.3 typing-extensions-4.15.0 virtualenv-21.2.0 2026-04-01T13:55:09.198 INFO:teuthology.orchestra.run.vm06.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-04-01T13:55:09.198 INFO:teuthology.orchestra.run.vm06.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/tox-venv/bin/python3 -m pip install --upgrade pip' command. 2026-04-01T13:55:09.241 INFO:teuthology.run_tasks:Running task tox... 2026-04-01T13:55:09.243 INFO:tasks.tox:Deploying tox from pip... 2026-04-01T13:55:09.244 DEBUG:teuthology.orchestra.run.vm06:> python3 -m venv /home/ubuntu/cephtest/tox-venv 2026-04-01T13:55:09.941 DEBUG:teuthology.orchestra.run.vm06:> source /home/ubuntu/cephtest/tox-venv/bin/activate && pip install tox 2026-04-01T13:55:10.098 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: tox in ./cephtest/tox-venv/lib/python3.9/site-packages (4.30.3) 2026-04-01T13:55:10.103 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: packaging>=25 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (26.0) 2026-04-01T13:55:10.104 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: typing-extensions>=4.14.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (4.15.0) 2026-04-01T13:55:10.104 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: filelock>=3.18 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (3.19.1) 2026-04-01T13:55:10.104 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: chardet>=5.2 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (5.2.0) 2026-04-01T13:55:10.104 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: colorama>=0.4.6 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (0.4.6) 2026-04-01T13:55:10.105 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: tomli>=2.2.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (2.4.1) 2026-04-01T13:55:10.105 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: pluggy>=1.6 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (1.6.0) 2026-04-01T13:55:10.105 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: platformdirs>=4.3.8 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (4.4.0) 2026-04-01T13:55:10.105 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: virtualenv>=20.31.2 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (21.2.0) 2026-04-01T13:55:10.106 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: pyproject-api>=1.9.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (1.9.1) 2026-04-01T13:55:10.106 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: cachetools>=6.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (6.2.6) 2026-04-01T13:55:10.128 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: distlib<1,>=0.3.7 in ./cephtest/tox-venv/lib/python3.9/site-packages (from virtualenv>=20.31.2->tox) (0.4.0) 2026-04-01T13:55:10.129 INFO:teuthology.orchestra.run.vm06.stdout:Requirement already satisfied: python-discovery>=1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from virtualenv>=20.31.2->tox) (1.2.1) 2026-04-01T13:55:10.146 INFO:teuthology.orchestra.run.vm06.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-04-01T13:55:10.146 INFO:teuthology.orchestra.run.vm06.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/tox-venv/bin/python3 -m pip install --upgrade pip' command. 2026-04-01T13:55:10.166 INFO:teuthology.run_tasks:Running task dedup-tests... 2026-04-01T13:55:10.169 DEBUG:tasks.dedup_tests:config is {'client.0': {'rgw_server': 'client.0'}} 2026-04-01T13:55:10.169 INFO:tasks.dedup_tests:Downloading dedup-tests... 2026-04-01T13:55:10.169 INFO:tasks.dedup_tests:Using branch tt-20.2.0-sse-s3-kmip-preview-not-for-production-2 from http://git.local/ceph.git for dedup tests 2026-04-01T13:55:10.169 DEBUG:teuthology.orchestra.run.vm06:> git clone -b tt-20.2.0-sse-s3-kmip-preview-not-for-production-2 http://git.local/ceph.git /home/ubuntu/cephtest/ceph 2026-04-01T13:55:10.185 INFO:teuthology.orchestra.run.vm06.stderr:Cloning into '/home/ubuntu/cephtest/ceph'... 2026-04-01T13:55:40.595 INFO:tasks.dedup_tests:Creating rgw user... 2026-04-01T13:55:40.595 DEBUG:tasks.dedup_tests:Creating user foo.client.0 on client.0 2026-04-01T13:55:40.595 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user create --uid foo.client.0 --display-name 'Mr. foo.client.0' --access-key TGTLZGBSBNIIRZKEOKZD --secret NA2iVSUlpQi/G1TXvPMGDH/aPsfv/f1iotQe5YQCqd6YqNeshIBRLA== --cluster ceph 2026-04-01T13:55:40.675 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T13:55:40.675 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T13:55:40.693 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.692+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.693+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.693+0000 7f6f0c12c900 20 realm 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.693+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.694+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.694+0000 7f6f0c12c900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.694+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.694+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.694 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.694+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.695 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.695+0000 7f6f0c12c900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:55:40.695 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.695+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.696 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.695+0000 7f6f0c12c900 20 rados_obj.operate() r=0 bl.length=1190 2026-04-01T13:55:40.696 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.695+0000 7f6f0c12c900 20 searching for the correct realm 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 RGWRados::pool_iterate: got zone_info.28f6e1de-cca5-4d4f-a2d4-c22f622f50fe 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 RGWRados::pool_iterate: got zonegroup_info.6d3bd279-f200-47e0-be3a-2ea7a6959538 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 RGWRados::pool_iterate: got default.zone. 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.704+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados_obj.operate() r=0 bl.length=46 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados_obj.operate() r=0 bl.length=470 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 zone default found 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 4 Realm: () 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 4 ZoneGroup: default (6d3bd279-f200-47e0-be3a-2ea7a6959538) 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 4 Zone: default (28f6e1de-cca5-4d4f-a2d4-c22f622f50fe) 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 zonegroup default 2026-04-01T13:55:40.705 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.705+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.706+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.706+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.706+0000 7f6f0c12c900 20 started sync module instance, tier type = 2026-04-01T13:55:40.706 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.706+0000 7f6f0c12c900 20 started zone id=28f6e1de-cca5-4d4f-a2d4-c22f622f50fe (name=default) with tier type = 2026-04-01T13:55:40.785 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.785+0000 7f6f0c12c900 20 add_watcher() i=5 2026-04-01T13:55:40.785 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.785+0000 7f6f0c12c900 20 add_watcher() i=0 2026-04-01T13:55:40.786 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 20 add_watcher() i=2 2026-04-01T13:55:40.786 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 20 add_watcher() i=6 2026-04-01T13:55:40.786 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 20 add_watcher() i=4 2026-04-01T13:55:40.786 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 20 add_watcher() i=7 2026-04-01T13:55:40.786 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 20 add_watcher() i=3 2026-04-01T13:55:40.787 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 20 add_watcher() i=1 2026-04-01T13:55:40.787 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.786+0000 7f6f0c12c900 2 all 8 watchers are set, enabling cache 2026-04-01T13:55:40.789 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.789+0000 7f6f0c12c900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-01T13:55:40.789 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.789+0000 7f6f0c12c900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-01T13:55:40.789 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.789+0000 7f6f0c12c900 5 note: GC not initialized 2026-04-01T13:55:40.789 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.789+0000 7f6eb4fe9640 20 reqs_thread_entry: start 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.842+0000 7f6f0c12c900 20 init_complete bucket index max shards: 11 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.842+0000 7f6f0c12c900 20 Filter name: none 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6eae7fc640 20 reqs_thread_entry: start 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 cache get: name=default.rgw.meta+users.uid+foo.client.0 : miss 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 cache put: name=default.rgw.meta+users.uid+foo.client.0 info.flags=0x0 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 adding default.rgw.meta+users.uid+foo.client.0 to cache LRU end 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 cache get: name=default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD : miss 2026-04-01T13:55:40.843 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 20 rados->read ofs=0 len=0 2026-04-01T13:55:40.844 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-01T13:55:40.844 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 cache put: name=default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD info.flags=0x0 2026-04-01T13:55:40.844 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 adding default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD to cache LRU end 2026-04-01T13:55:40.844 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 cache get: name=default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD : hit (negative entry) 2026-04-01T13:55:40.844 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.843+0000 7f6f0c12c900 10 cache get: name=default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD : hit (negative entry) 2026-04-01T13:55:40.845 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.845+0000 7f6f0c12c900 10 cache put: name=default.rgw.meta+users.uid+foo.client.0 info.flags=0x17 2026-04-01T13:55:40.845 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.845+0000 7f6f0c12c900 10 moving default.rgw.meta+users.uid+foo.client.0 to cache LRU end 2026-04-01T13:55:40.845 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.845+0000 7f6f0c12c900 10 distributing notification oid=default.rgw.control:notify.0 cni=[op: 0, obj: default.rgw.meta:users.uid:foo.client.0, ofs0, ns] 2026-04-01T13:55:40.846 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.845+0000 7f6ee2ffd640 10 rgw watcher librados: RGWWatcher::handle_notify() notify_id 154618822656 cookie 94597104558016 notifier 4691 bl.length()=628 2026-04-01T13:55:40.846 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.845+0000 7f6ee2ffd640 10 rgw watcher librados: cache put: name=default.rgw.meta+users.uid+foo.client.0 info.flags=0x17 2026-04-01T13:55:40.846 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.845+0000 7f6ee2ffd640 10 rgw watcher librados: moving default.rgw.meta+users.uid+foo.client.0 to cache LRU end 2026-04-01T13:55:40.847 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.847+0000 7f6f0c12c900 10 cache put: name=default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD info.flags=0x7 2026-04-01T13:55:40.847 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.847+0000 7f6f0c12c900 10 moving default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD to cache LRU end 2026-04-01T13:55:40.847 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.847+0000 7f6f0c12c900 10 distributing notification oid=default.rgw.control:notify.0 cni=[op: 0, obj: default.rgw.meta:users.keys:TGTLZGBSBNIIRZKEOKZD, ofs0, ns] 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.847+0000 7f6ee2ffd640 10 rgw watcher librados: RGWWatcher::handle_notify() notify_id 154618822657 cookie 94597104558016 notifier 4691 bl.length()=186 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.847+0000 7f6ee2ffd640 10 rgw watcher librados: cache put: name=default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD info.flags=0x7 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.847+0000 7f6ee2ffd640 10 rgw watcher librados: moving default.rgw.meta+users.keys+TGTLZGBSBNIIRZKEOKZD to cache LRU end 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "user_id": "foo.client.0", 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "display_name": "Mr. foo.client.0", 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "email": "", 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "suspended": 0, 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "max_buckets": 1000, 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "subusers": [], 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "keys": [ 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: { 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "user": "foo.client.0", 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "access_key": "TGTLZGBSBNIIRZKEOKZD", 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "secret_key": "NA2iVSUlpQi/G1TXvPMGDH/aPsfv/f1iotQe5YQCqd6YqNeshIBRLA==", 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "active": true, 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "create_date": "2026-04-01T13:55:40.844922Z" 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: } 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: ], 2026-04-01T13:55:40.848 INFO:teuthology.orchestra.run.vm06.stdout: "swift_keys": [], 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "caps": [], 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "op_mask": "read, write, delete", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "default_placement": "", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "default_storage_class": "", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "placement_tags": [], 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "bucket_quota": { 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "enabled": false, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "check_on_raw": false, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "max_size": -1, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "max_size_kb": 0, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "max_objects": -1 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: }, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "user_quota": { 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "enabled": false, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "check_on_raw": false, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "max_size": -1, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "max_size_kb": 0, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "max_objects": -1 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: }, 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "temp_url_keys": [], 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "type": "rgw", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "mfa_ids": [], 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "account_id": "", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "path": "/", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "create_date": "2026-04-01T13:55:40.844915Z", 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "tags": [], 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: "group_ids": [] 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-04-01T13:55:40.849 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:55:40.852 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=2 2026-04-01T13:55:40.852 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 2 removed watcher, disabling cache 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=4 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=0 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=5 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=6 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=3 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=7 2026-04-01T13:55:40.853 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T13:55:40.852+0000 7f6f0c12c900 20 remove_watcher() i=1 2026-04-01T13:55:40.859 INFO:tasks.dedup_tests:Configuring dedup-tests... 2026-04-01T13:55:40.860 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-01T13:55:40.860 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/deduptests.client.0.conf 2026-04-01T13:55:40.916 INFO:tasks.dedup_tests:Running dedup-tests... 2026-04-01T13:55:40.916 DEBUG:teuthology.orchestra.run.vm06:dedup tests against rgw> source /home/ubuntu/cephtest/tox-venv/bin/activate && cd /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/ && DEDUPTESTS_CONF=./deduptests.client.0.conf tox -- -v -m 'basic_test or request_test or example_test' 2026-04-01T13:55:41.290 INFO:teuthology.orchestra.run.vm06.stdout:py: install_deps> python -I -m pip install -r requirements.txt 2026-04-01T13:55:43.827 INFO:teuthology.orchestra.run.vm06.stdout:py: commands[0]> pytest -v -m 'basic_test or request_test or example_test' 2026-04-01T13:55:43.915 INFO:teuthology.orchestra.run.vm06.stdout:============================= test session starts ============================== 2026-04-01T13:55:43.915 INFO:teuthology.orchestra.run.vm06.stdout:platform linux -- Python 3.9.23, pytest-8.4.2, pluggy-1.6.0 -- /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/bin/python 2026-04-01T13:55:43.915 INFO:teuthology.orchestra.run.vm06.stdout:cachedir: .tox/py/.pytest_cache 2026-04-01T13:55:43.915 INFO:teuthology.orchestra.run.vm06.stdout:rootdir: /home/ubuntu/cephtest/ceph/src/test/rgw/dedup 2026-04-01T13:55:43.915 INFO:teuthology.orchestra.run.vm06.stdout:configfile: pytest.ini 2026-04-01T13:55:44.011 INFO:teuthology.orchestra.run.vm06.stdout:collecting ... collected 34 items 2026-04-01T13:55:44.011 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T13:55:44.135 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_etag_corruption PASSED [ 2%] 2026-04-01T13:55:44.135 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_md5_collisions PASSED [ 5%] 2026-04-01T13:55:44.135 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_small PASSED [ 8%] 2026-04-01T13:55:44.136 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_small_with_tenants PASSED [ 11%] 2026-04-01T13:55:44.136 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_0_with_tenants PASSED [ 14%] 2026-04-01T13:55:44.136 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_0 PASSED [ 17%] 2026-04-01T13:55:44.137 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_1_with_tenants PASSED [ 20%] 2026-04-01T13:55:44.137 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_1 PASSED [ 23%] 2026-04-01T13:55:44.137 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_2_with_tenants PASSED [ 26%] 2026-04-01T13:55:44.138 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_2 PASSED [ 29%] 2026-04-01T13:55:44.138 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_with_remove_multi_tenants PASSED [ 32%] 2026-04-01T13:55:44.138 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_with_remove PASSED [ 35%] 2026-04-01T13:55:44.139 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_multipart_with_tenants PASSED [ 38%] 2026-04-01T13:55:44.139 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_multipart PASSED [ 41%] 2026-04-01T13:55:44.139 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_basic_with_tenants PASSED [ 44%] 2026-04-01T13:55:44.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_basic PASSED [ 47%] 2026-04-01T13:55:44.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_small_multipart_with_tenants PASSED [ 50%] 2026-04-01T13:55:44.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_small_multipart PASSED [ 52%] 2026-04-01T13:55:44.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_large_scale_with_tenants PASSED [ 55%] 2026-04-01T13:55:44.141 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_large_scale PASSED [ 58%] 2026-04-01T13:55:44.141 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_empty_bucket PASSED [ 61%] 2026-04-01T13:55:44.141 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_inc_loop_with_tenants PASSED [ 64%] 2026-04-01T13:55:50.503 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_with_tenants 2026-04-01T13:55:50.503 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:55:50.503 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:55:51.073 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 67%] 2026-04-01T13:57:53.694 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_multipart 2026-04-01T13:57:53.694 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:57:53.694 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:57:57.210 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 70%] 2026-04-01T13:58:07.955 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_basic 2026-04-01T13:58:07.955 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:58:07.955 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:58:08.471 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 73%] 2026-04-01T13:58:19.174 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_multipart 2026-04-01T13:58:19.174 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:58:19.174 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:58:19.748 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 76%] 2026-04-01T13:58:25.572 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small 2026-04-01T13:58:25.572 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:58:25.572 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:58:26.013 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 79%] 2026-04-01T13:58:41.430 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_large_mix 2026-04-01T13:58:41.430 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:58:41.430 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:58:42.636 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 82%] 2026-04-01T13:59:01.549 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_basic_with_tenants 2026-04-01T13:59:01.549 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T13:59:01.549 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T13:59:02.538 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 85%] 2026-04-01T14:00:11.206 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_multipart_with_tenants 2026-04-01T14:00:11.206 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T14:00:11.206 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T14:00:13.282 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 88%] 2026-04-01T14:00:22.760 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_multipart_with_tenants 2026-04-01T14:00:22.760 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T14:00:22.760 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T14:00:24.283 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 91%] 2026-04-01T14:05:28.133 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:05:28.132+0000 7fa63aeaa640 -1 log_channel(cluster) log [ERR] : Health check failed: mon c is very low on available space (MON_DISK_CRIT) 2026-04-01T14:05:33.432 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:05:33.431+0000 7fa63d6af640 -1 log_channel(cluster) log [ERR] : Health check update: mons a,c are very low on available space (MON_DISK_CRIT) 2026-04-01T14:07:42.431 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:07:42.429+0000 7fa63d6af640 -1 log_channel(cluster) log [ERR] : Health check update: mons a,b,c are very low on available space (MON_DISK_CRIT) 2026-04-01T14:07:59.124 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_large_scale_with_tenants 2026-04-01T14:07:59.124 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T14:07:59.124 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-01T14:07:59.124 INFO:teuthology.orchestra.run.vm06.stdout:INFO dedup.test_dedup:test_dedup.py:1288 [64] obj_count=65313, upload=438(sec), exec=5(sec), verify=0(sec) 2026-04-01T14:10:00.000 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:10:00.000+0000 7fa63d6af640 -1 log_channel(cluster) log [ERR] : overall HEALTH_ERR mons a,b,c are very low on available space 2026-04-01T14:10:12.214 INFO:teuthology.orchestra.run.vm06.stdout:PASSED [ 94%] 2026-04-01T14:10:45.361 INFO:tasks.ceph.osd.0.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.0.log: (28) No space left on device 2026-04-01T14:10:45.362 INFO:tasks.rgw.client.0.vm06.stdout:problem writing to /var/log/ceph/rgw.ceph.client.0.log: (28) No space left on device 2026-04-01T14:10:45.362 INFO:tasks.ceph.osd.1.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.1.log: (28) No space left on device 2026-04-01T14:10:45.399 INFO:tasks.ceph.osd.2.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.2.log: (28) No space left on device 2026-04-01T14:10:45.399 INFO:tasks.ceph.osd.3.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.3.log: (28) No space left on device 2026-04-01T14:10:45.400 INFO:tasks.ceph.osd.1.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.1.log: (28) No space left on device 2026-04-01T14:10:45.411 INFO:tasks.ceph.osd.0.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.0.log: (28) No space left on device 2026-04-01T14:10:45.485 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:10:45.645 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:10:47.881 INFO:tasks.ceph.mgr.y.vm06.stderr:problem writing to /var/log/ceph/ceph-mgr.y.log: (28) No space left on device 2026-04-01T14:11:27.528 INFO:tasks.ceph.osd.6.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.6.log: (28) No space left on device 2026-04-01T14:11:27.529 INFO:tasks.ceph.osd.5.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.5.log: (28) No space left on device 2026-04-01T14:11:27.529 INFO:tasks.ceph.osd.5.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.5.log: (28) No space left on device 2026-04-01T14:11:27.529 INFO:tasks.ceph.osd.7.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.7.log: (28) No space left on device 2026-04-01T14:11:27.530 INFO:tasks.ceph.osd.5.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.5.log: (28) No space left on device 2026-04-01T14:11:27.534 INFO:tasks.ceph.osd.7.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.7.log: (28) No space left on device 2026-04-01T14:11:27.534 INFO:tasks.ceph.osd.4.vm08.stderr:problem writing to /var/log/ceph/ceph-osd.4.log: (28) No space left on device 2026-04-01T14:11:27.653 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:27.885 INFO:tasks.ceph.mgr.x.vm08.stderr:problem writing to /var/log/ceph/ceph-mgr.x.log: (28) No space left on device 2026-04-01T14:11:29.853 INFO:tasks.ceph.mon.b.vm08.stderr:2026-04-01T14:11:29.852+0000 7fc8e778d640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-b/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:29.853 INFO:tasks.ceph.mon.b.vm08.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fc8e778d640 time 2026-04-01T14:11:29.854855+0000 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fc8efd901fd] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 2: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 9: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr:2026-04-01T14:11:29.854+0000 7fc8e778d640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fc8e778d640 time 2026-04-01T14:11:29.854855+0000 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fc8efd901fd] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 2: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 9: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr:*** Caught signal (Aborted) ** 2026-04-01T14:11:29.855 INFO:tasks.ceph.mon.b.vm08.stderr: in thread 7fc8e778d640 thread_name:ms_dispatch 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fc8eee3fc30] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fc8eee8d02c] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 3: raise() 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 4: abort() 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fc8efd902ba] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 6: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 13: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr:2026-04-01T14:11:29.855+0000 7fc8e778d640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: in thread 7fc8e778d640 thread_name:ms_dispatch 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fc8eee3fc30] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fc8eee8d02c] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 3: raise() 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 4: abort() 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fc8efd902ba] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 6: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 13: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:29.856 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.857 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: -2> 2026-04-01T14:11:29.852+0000 7fc8e778d640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-b/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: -1> 2026-04-01T14:11:29.854+0000 7fc8e778d640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fc8e778d640 time 2026-04-01T14:11:29.854855+0000 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fc8efd901fd] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 2: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 9: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 0> 2026-04-01T14:11:29.855+0000 7fc8e778d640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: in thread 7fc8e778d640 thread_name:ms_dispatch 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fc8eee3fc30] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fc8eee8d02c] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 3: raise() 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 4: abort() 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fc8efd902ba] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 6: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 13: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.866 INFO:tasks.ceph.mon.b.vm08.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.867 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.869 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: -9999> 2026-04-01T14:11:29.852+0000 7fc8e778d640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-b/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: -9998> 2026-04-01T14:11:29.854+0000 7fc8e778d640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fc8e778d640 time 2026-04-01T14:11:29.854855+0000 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fc8efd901fd] 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 2: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.870 INFO:tasks.ceph.mon.b.vm08.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 9: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: -9997> 2026-04-01T14:11:29.855+0000 7fc8e778d640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: in thread 7fc8e778d640 thread_name:ms_dispatch 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fc8eee3fc30] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fc8eee8d02c] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 3: raise() 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 4: abort() 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fc8efd902ba] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 6: ceph-mon(+0x2a61ac) [0x5651e98c61ac] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5651e99a9865] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x5651e99b2901] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x5651e99aecd0] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x5651e99af887] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x5651e991ee3d] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x5651e99135c6] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 13: ceph-mon(+0x2b333c) [0x5651e98d333c] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7fc8f0006848] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7fc8f009ac51] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7fc8eee8b2ea] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7fc8eef103c0] 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:29.871 INFO:tasks.ceph.mon.b.vm08.stderr: 2026-04-01T14:11:30.053 INFO:tasks.ceph.mon.b.vm08.stderr:daemon-helper: command crashed with signal 6 2026-04-01T14:11:30.366 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:11:30.366+0000 7fa63d6af640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-a/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:30.366 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = '1151' value size = 13517) 2026-04-01T14:11:30.366 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = 'pending_v' value size = 8) 2026-04-01T14:11:30.366 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = 'pending_pn' value size = 8) 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa63d6af640 time 2026-04-01T14:11:30.366971+0000 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa6435901fd] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 4: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 6: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 7: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr:*** Caught signal (Aborted) ** 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: in thread 7fa63d6af640 thread_name:safe_timer 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:11:30.367+0000 7fa63d6af640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa63d6af640 time 2026-04-01T14:11:30.366971+0000 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa6435901fd] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 4: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 6: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 7: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.367 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa64263fc30] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa64268d02c] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 3: raise() 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 4: abort() 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa6435902ba] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 9: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 10: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 11: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr:2026-04-01T14:11:30.368+0000 7fa63d6af640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: in thread 7fa63d6af640 thread_name:safe_timer 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa64263fc30] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa64268d02c] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 3: raise() 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 4: abort() 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa6435902ba] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 9: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 10: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 11: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:30.368 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.369 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: -2> 2026-04-01T14:11:30.366+0000 7fa63d6af640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-a/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = '1151' value size = 13517) 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = 'pending_v' value size = 8) 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = 'pending_pn' value size = 8) 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: -1> 2026-04-01T14:11:30.367+0000 7fa63d6af640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa63d6af640 time 2026-04-01T14:11:30.366971+0000 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa6435901fd] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 4: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 6: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 7: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 0> 2026-04-01T14:11:30.368+0000 7fa63d6af640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: in thread 7fa63d6af640 thread_name:safe_timer 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa64263fc30] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa64268d02c] 2026-04-01T14:11:30.390 INFO:tasks.ceph.mon.a.vm06.stderr: 3: raise() 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 4: abort() 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa6435902ba] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 9: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 10: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 11: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.391 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.392 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.393 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.394 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.395 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.396 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.397 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.398 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.399 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.399 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.399 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.399 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.399 INFO:tasks.ceph.mon.a.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: -9999> 2026-04-01T14:11:30.366+0000 7fa63d6af640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-a/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = '1151' value size = 13517) 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = 'pending_v' value size = 8) 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr:PutCF( prefix = paxos key = 'pending_pn' value size = 8) 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: -9998> 2026-04-01T14:11:30.367+0000 7fa63d6af640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa63d6af640 time 2026-04-01T14:11:30.366971+0000 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa6435901fd] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 4: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 6: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 7: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.402 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: -9997> 2026-04-01T14:11:30.368+0000 7fa63d6af640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: in thread 7fa63d6af640 thread_name:safe_timer 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa64263fc30] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa64268d02c] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 3: raise() 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 4: abort() 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa6435902ba] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55829a7bf1ac] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x55829a93d95c] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 8: (Paxos::propose_pending()+0x11b) [0x55829a94b70b] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 9: (Paxos::trigger_propose()+0x118) [0x55829a94bb08] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 10: (PaxosService::propose_pending()+0x176) [0x55829a94be46] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 11: ceph-mon(+0x2a644d) [0x55829a7bf44d] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7fa6436dc550] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa6436dcfb1] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7fa64268b2ea] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7fa6427103c0] 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:30.403 INFO:tasks.ceph.mon.a.vm06.stderr: 2026-04-01T14:11:30.501 INFO:tasks.ceph.mon.a.vm06.stderr:daemon-helper: command crashed with signal 6 2026-04-01T14:11:32.806 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~0s 2026-04-01T14:11:32.806 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~0s 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr:2026-04-01T14:11:35.654+0000 7fa7e3d48640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-c/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa7e3d48640 time 2026-04-01T14:11:35.654152+0000 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa7e9b901fd] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr:2026-04-01T14:11:35.654+0000 7fa7e3d48640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa7e3d48640 time 2026-04-01T14:11:35.654152+0000 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa7e9b901fd] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.654 INFO:tasks.ceph.mon.c.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr:*** Caught signal (Aborted) ** 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: in thread 7fa7e3d48640 thread_name:safe_timer 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa7e8c3fc30] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa7e8c8d02c] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 3: raise() 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 4: abort() 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa7e9b902ba] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 9: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr:2026-04-01T14:11:35.655+0000 7fa7e3d48640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: in thread 7fa7e3d48640 thread_name:safe_timer 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa7e8c3fc30] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa7e8c8d02c] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 3: raise() 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 4: abort() 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa7e9b902ba] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 9: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.655 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.676 INFO:tasks.ceph.mon.c.vm06.stderr: -2> 2026-04-01T14:11:35.654+0000 7fa7e3d48640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-c/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:35.676 INFO:tasks.ceph.mon.c.vm06.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: -1> 2026-04-01T14:11:35.654+0000 7fa7e3d48640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa7e3d48640 time 2026-04-01T14:11:35.654152+0000 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa7e9b901fd] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 0> 2026-04-01T14:11:35.655+0000 7fa7e3d48640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: in thread 7fa7e3d48640 thread_name:safe_timer 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa7e8c3fc30] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa7e8c8d02c] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 3: raise() 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 4: abort() 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa7e9b902ba] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 9: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.677 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.678 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.679 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.680 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: -9999> 2026-04-01T14:11:35.654+0000 7fa7e3d48640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-c/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: -9998> 2026-04-01T14:11:35.654+0000 7fa7e3d48640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fa7e3d48640 time 2026-04-01T14:11:35.654152+0000 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-9-ge2ce8426bc5/BUILD/ceph-20.2.0-9-ge2ce8426bc5/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fa7e9b901fd] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: -9997> 2026-04-01T14:11:35.655+0000 7fa7e3d48640 -1 *** Caught signal (Aborted) ** 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: in thread 7fa7e3d48640 thread_name:safe_timer 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: ceph version 20.2.0-9-ge2ce8426bc5 (e2ce8426bc5387b913d1b06bd31280ced108bbb6) tentacle (stable - RelWithDebInfo) 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fa7e8c3fc30] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fa7e8c8d02c] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 3: raise() 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 4: abort() 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fa7e9b902ba] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x5606f9c191ac] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x5606f9cfc865] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x5606f9d060c1] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 9: (Elector::dead_ping(int)+0x1a1) [0x5606f9cfdb91] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x5606f9c1944d] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fa7e9cdc550] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fa7e9cdcfb1] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fa7e8c8b2ea] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fa7e8d103c0] 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-01T14:11:35.681 INFO:tasks.ceph.mon.c.vm06.stderr: 2026-04-01T14:11:35.735 INFO:tasks.ceph.mon.c.vm06.stderr:daemon-helper: command crashed with signal 6 2026-04-01T14:11:39.113 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~6s 2026-04-01T14:11:39.113 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~0s 2026-04-01T14:11:39.113 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~6s 2026-04-01T14:11:45.175 INFO:tasks.rgw.client.1.vm08.stdout:problem writing to /var/log/ceph/rgw.ceph.client.1.log: tee: /var/log/ceph/rgw.ceph.client.1.stdout: No space left on device 2026-04-01T14:11:45.175 INFO:tasks.rgw.client.1.vm08.stdout:(28) No space left on device 2026-04-01T14:11:45.419 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~13s 2026-04-01T14:11:45.419 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~6s 2026-04-01T14:11:45.419 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~13s 2026-04-01T14:11:51.725 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~19s 2026-04-01T14:11:51.725 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~13s 2026-04-01T14:11:51.725 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~19s 2026-04-01T14:11:58.030 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~25s 2026-04-01T14:11:58.030 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~19s 2026-04-01T14:11:58.030 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~25s 2026-04-01T14:12:04.335 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~32s 2026-04-01T14:12:04.335 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~25s 2026-04-01T14:12:04.335 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~32s 2026-04-01T14:12:10.640 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~38s 2026-04-01T14:12:10.640 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~32s 2026-04-01T14:12:10.640 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~38s 2026-04-01T14:12:16.944 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~44s 2026-04-01T14:12:16.945 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~38s 2026-04-01T14:12:16.945 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~44s 2026-04-01T14:12:23.251 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~51s 2026-04-01T14:12:23.252 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~44s 2026-04-01T14:12:23.252 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~51s 2026-04-01T14:12:29.557 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~57s 2026-04-01T14:12:29.557 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~50s 2026-04-01T14:12:29.557 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~57s 2026-04-01T14:12:35.862 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~63s 2026-04-01T14:12:35.862 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~57s 2026-04-01T14:12:35.862 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~63s 2026-04-01T14:12:42.167 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~69s 2026-04-01T14:12:42.167 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~63s 2026-04-01T14:12:42.167 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~69s 2026-04-01T14:12:48.472 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~76s 2026-04-01T14:12:48.472 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~69s 2026-04-01T14:12:48.472 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~76s 2026-04-01T14:12:54.777 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~82s 2026-04-01T14:12:54.777 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~76s 2026-04-01T14:12:54.777 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~82s 2026-04-01T14:13:01.083 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~88s 2026-04-01T14:13:01.084 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~82s 2026-04-01T14:13:01.084 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~88s 2026-04-01T14:13:07.390 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~95s 2026-04-01T14:13:07.390 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~88s 2026-04-01T14:13:07.390 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~95s 2026-04-01T14:13:13.695 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~101s 2026-04-01T14:13:13.695 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~95s 2026-04-01T14:13:13.695 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~101s 2026-04-01T14:13:20.000 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~107s 2026-04-01T14:13:20.000 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~101s 2026-04-01T14:13:20.000 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~107s 2026-04-01T14:13:26.306 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~114s 2026-04-01T14:13:26.306 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~107s 2026-04-01T14:13:26.306 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~114s 2026-04-01T14:13:32.613 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~120s 2026-04-01T14:13:32.614 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~114s 2026-04-01T14:13:32.614 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~120s 2026-04-01T14:13:38.921 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~126s 2026-04-01T14:13:38.921 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~120s 2026-04-01T14:13:38.921 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~126s 2026-04-01T14:13:45.227 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~133s 2026-04-01T14:13:45.227 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~126s 2026-04-01T14:13:45.227 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~133s 2026-04-01T14:13:51.533 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~139s 2026-04-01T14:13:51.533 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~132s 2026-04-01T14:13:51.533 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~139s 2026-04-01T14:13:57.840 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~145s 2026-04-01T14:13:57.840 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~139s 2026-04-01T14:13:57.840 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~145s 2026-04-01T14:14:04.145 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~151s 2026-04-01T14:14:04.146 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~145s 2026-04-01T14:14:04.146 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~151s 2026-04-01T14:14:10.451 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~158s 2026-04-01T14:14:10.451 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~151s 2026-04-01T14:14:10.451 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~158s 2026-04-01T14:14:16.757 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~164s 2026-04-01T14:14:16.757 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~158s 2026-04-01T14:14:16.757 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~164s 2026-04-01T14:14:23.064 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~170s 2026-04-01T14:14:23.064 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~164s 2026-04-01T14:14:23.064 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~170s 2026-04-01T14:14:29.369 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~177s 2026-04-01T14:14:29.369 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~170s 2026-04-01T14:14:29.369 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~177s 2026-04-01T14:14:35.674 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~183s 2026-04-01T14:14:35.674 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~177s 2026-04-01T14:14:35.674 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~183s 2026-04-01T14:14:41.980 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~189s 2026-04-01T14:14:41.980 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~183s 2026-04-01T14:14:41.980 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~189s 2026-04-01T14:14:48.285 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~196s 2026-04-01T14:14:48.285 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~189s 2026-04-01T14:14:48.285 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~196s 2026-04-01T14:14:54.592 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~202s 2026-04-01T14:14:54.592 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~195s 2026-04-01T14:14:54.592 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~202s 2026-04-01T14:15:00.898 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~208s 2026-04-01T14:15:00.898 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~202s 2026-04-01T14:15:00.898 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~208s 2026-04-01T14:15:07.204 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~215s 2026-04-01T14:15:07.204 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~208s 2026-04-01T14:15:07.204 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~215s 2026-04-01T14:15:13.514 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~221s 2026-04-01T14:15:13.514 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~214s 2026-04-01T14:15:13.514 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~221s 2026-04-01T14:15:19.820 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~227s 2026-04-01T14:15:19.820 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~221s 2026-04-01T14:15:19.820 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~227s 2026-04-01T14:15:26.126 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~233s 2026-04-01T14:15:26.127 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~227s 2026-04-01T14:15:26.127 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~233s 2026-04-01T14:15:32.433 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~240s 2026-04-01T14:15:32.433 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~233s 2026-04-01T14:15:32.433 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~240s 2026-04-01T14:15:38.740 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~246s 2026-04-01T14:15:38.740 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~240s 2026-04-01T14:15:38.740 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~246s 2026-04-01T14:15:45.046 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~252s 2026-04-01T14:15:45.046 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~246s 2026-04-01T14:15:45.047 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~252s 2026-04-01T14:15:51.351 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~259s 2026-04-01T14:15:51.351 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~252s 2026-04-01T14:15:51.351 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~259s 2026-04-01T14:15:57.658 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~265s 2026-04-01T14:15:57.658 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~259s 2026-04-01T14:15:57.658 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~265s 2026-04-01T14:16:03.967 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~271s 2026-04-01T14:16:03.967 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~265s 2026-04-01T14:16:03.967 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~271s 2026-04-01T14:16:10.273 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~278s 2026-04-01T14:16:10.273 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~271s 2026-04-01T14:16:10.273 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~278s 2026-04-01T14:16:16.581 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~284s 2026-04-01T14:16:16.582 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~277s 2026-04-01T14:16:16.582 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~284s 2026-04-01T14:16:22.887 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~290s 2026-04-01T14:16:22.887 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~284s 2026-04-01T14:16:22.887 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~290s 2026-04-01T14:16:29.196 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~296s 2026-04-01T14:16:29.196 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~290s 2026-04-01T14:16:29.196 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~296s 2026-04-01T14:16:35.504 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~303s 2026-04-01T14:16:35.504 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~296s 2026-04-01T14:16:35.504 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~303s 2026-04-01T14:16:35.504 INFO:tasks.daemonwatchdog.daemon_watchdog:BARK! unmounting mounts and killing all daemons 2026-04-01T14:16:36.812 INFO:tasks.ceph.osd.0:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.1:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.2:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.3:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.4:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.5:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.6:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.7:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.rgw.client.0:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.rgw.client.1:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.rgw.client.2:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.mgr.y:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.mgr.x:Sent signal 15 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T14:16:36.813+0000 7f68f0c21640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 5 (PID: 57879) UID: 0 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T14:16:36.813+0000 7f68f0c21640 -1 osd.5 69 *** Got signal Terminated *** 2026-04-01T14:16:36.813 INFO:tasks.ceph.osd.5.vm08.stderr:2026-04-01T14:16:36.813+0000 7f68f0c21640 -1 osd.5 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T14:16:36.812+0000 7fe857e8c640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 1 (PID: 64987) UID: 0 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T14:16:36.812+0000 7fe857e8c640 -1 osd.1 69 *** Got signal Terminated *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.1.vm06.stderr:2026-04-01T14:16:36.812+0000 7fe857e8c640 -1 osd.1 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T14:16:36.813+0000 7fee6644d640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 6 (PID: 57875) UID: 0 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T14:16:36.813+0000 7fee6644d640 -1 osd.6 69 *** Got signal Terminated *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.6.vm08.stderr:2026-04-01T14:16:36.813+0000 7fee6644d640 -1 osd.6 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T14:16:36.813+0000 7fc64c25c640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 7 (PID: 57880) UID: 0 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T14:16:36.813+0000 7fc64c25c640 -1 osd.7 69 *** Got signal Terminated *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.7.vm08.stderr:2026-04-01T14:16:36.813+0000 7fc64c25c640 -1 osd.7 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T14:16:36.813+0000 7f33106c4640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 4 (PID: 57870) UID: 0 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T14:16:36.813+0000 7f33106c4640 -1 osd.4 69 *** Got signal Terminated *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.4.vm08.stderr:2026-04-01T14:16:36.813+0000 7f33106c4640 -1 osd.4 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T14:16:36.812+0000 7f7b6a51d640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 0 (PID: 64986) UID: 0 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T14:16:36.812+0000 7f7b6a51d640 -1 osd.0 69 *** Got signal Terminated *** 2026-04-01T14:16:36.814 INFO:tasks.ceph.osd.0.vm06.stderr:2026-04-01T14:16:36.812+0000 7f7b6a51d640 -1 osd.0 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.814 INFO:tasks.rgw.client.1.vm08.stdout:2026-04-01T14:16:36.813+0000 7fb64c1ec640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper term radosgw --rgw-frontends beast port=80 -n client.1 --cluster ceph -k /etc/ceph/ceph.client.1.keyring --log-file /var/log/ceph/rgw.ceph.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.1.sock --foreground (PID: 62159) UID: 0 2026-04-01T14:16:36.815 INFO:tasks.rgw.client.2.vm09.stdout:2026-04-01T14:16:36.814+0000 7f7bc3bdb640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper term radosgw --rgw-frontends beast port=80 -n client.2 --cluster ceph -k /etc/ceph/ceph.client.2.keyring --log-file /var/log/ceph/rgw.ceph.client.2.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.2.sock --foreground (PID: 51855) UID: 0 2026-04-01T14:16:36.815 INFO:tasks.rgw.client.2.vm09.stdout:2026-04-01T14:16:36.814+0000 7f7bc9323980 -1 shutting down 2026-04-01T14:16:36.815 INFO:tasks.rgw.client.1.vm08.stdout:2026-04-01T14:16:36.814+0000 7fb64fa4d980 -1 shutting down 2026-04-01T14:16:36.815 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T14:16:36.813+0000 7f8cb6fe4640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 3 (PID: 64985) UID: 0 2026-04-01T14:16:36.815 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T14:16:36.813+0000 7f8cb6fe4640 -1 osd.3 69 *** Got signal Terminated *** 2026-04-01T14:16:36.815 INFO:tasks.ceph.osd.3.vm06.stderr:2026-04-01T14:16:36.813+0000 7f8cb6fe4640 -1 osd.3 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.821 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T14:16:36.820+0000 7fc9e5b96640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2 (PID: 64975) UID: 0 2026-04-01T14:16:36.822 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T14:16:36.820+0000 7fc9e5b96640 -1 osd.2 69 *** Got signal Terminated *** 2026-04-01T14:16:36.822 INFO:tasks.ceph.osd.2.vm06.stderr:2026-04-01T14:16:36.820+0000 7fc9e5b96640 -1 osd.2 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-01T14:16:36.822 INFO:tasks.rgw.client.0.vm06.stdout:2026-04-01T14:16:36.820+0000 7f56b7c7c640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper term radosgw --rgw-frontends beast port=80 -n client.0 --cluster ceph -k /etc/ceph/ceph.client.0.keyring --log-file /var/log/ceph/rgw.ceph.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock --foreground (PID: 71366) UID: 0 2026-04-01T14:16:36.822 INFO:tasks.rgw.client.0.vm06.stdout:2026-04-01T14:16:36.820+0000 7f56bd323980 -1 shutting down 2026-04-01T14:16:37.014 INFO:tasks.ceph.mgr.y.vm06.stderr:daemon-helper: command crashed with signal 15 2026-04-01T14:17:07.124 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_large_scale 2026-04-01T14:17:07.124 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------- live log call --------------------------------- 2026-04-01T14:17:07.124 INFO:teuthology.orchestra.run.vm06.stdout:WARNING dedup.test_dedup:test_dedup.py:2748 test_dedup_dry_large_scale: failed!! 2026-04-01T14:17:13.872 INFO:teuthology.orchestra.run.vm06.stdout:FAILED [ 97%] 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_cleanup PASSED [100%] 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:=================================== FAILURES =================================== 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:__________________________ test_dedup_dry_large_scale __________________________ 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:self = 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: def _new_conn(self): 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: """Establish a socket connection and set nodelay settings on it. 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: :return: New socket connection. 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: """ 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: extra_kw = {} 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: if self.source_address: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: extra_kw["source_address"] = self.source_address 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: if self.socket_options: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: extra_kw["socket_options"] = self.socket_options 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:> conn = connection.create_connection( 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: (self._dns_host, self.port), self.timeout, **extra_kw 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:174: 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py:95: in create_connection 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: raise err 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-01T14:17:13.876 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout:address = ('vm06.local', 80), timeout = 60, source_address = None 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout:socket_options = [(6, 1, 1)] 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: def create_connection( 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: address, 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: timeout=socket._GLOBAL_DEFAULT_TIMEOUT, 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: source_address=None, 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: socket_options=None, 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: ): 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: """Connect to *address* and return the socket object. 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: Convenience function. Connect to *address* (a 2-tuple ``(host, 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: port)``) and return the socket object. Passing the optional 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: *timeout* parameter will set the timeout on the socket instance 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: before attempting to connect. If no *timeout* is supplied, the 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: global default timeout setting returned by :func:`socket.getdefaulttimeout` 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: is used. If *source_address* is set it must be a tuple of (host, port) 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: for the socket to bind as a source address before making the connection. 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: An host of '' or port 0 tells the OS to use the default. 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: """ 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: host, port = address 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: if host.startswith("["): 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: host = host.strip("[]") 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: err = None 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: # Using the value from allowed_gai_family() in the context of getaddrinfo lets 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: # us select whether to work with IPv4 DNS records, IPv6 records, or both. 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: # The original create_connection function always returns all records. 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: family = allowed_gai_family() 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: host.encode("idna") 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: except UnicodeError: 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: return six.raise_from( 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: LocationParseError(u"'%s', label empty or too long" % host), None 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.877 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: af, socktype, proto, canonname, sa = res 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: sock = None 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: sock = socket.socket(af, socktype, proto) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: # If provided, set socket level options before connecting. 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: _set_socket_options(sock, socket_options) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: sock.settimeout(timeout) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: if source_address: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: sock.bind(source_address) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:> sock.connect(sa) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:E ConnectionRefusedError: [Errno 111] Connection refused 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py:85: ConnectionRefusedError 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:During handling of the above exception, another exception occurred: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:self = 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:request = 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: def send(self, request): 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: proxy_url = self._proxy_config.proxy_url_for(request.url) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: manager = self._get_connection_manager(request.url, proxy_url) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: conn = manager.connection_from_url(request.url) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: self._setup_ssl_cert(conn, request.url, self._verify) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: if ensure_boolean( 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: os.environ.get('BOTO_EXPERIMENTAL__ADD_PROXY_HOST_HEADER', '') 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: ): 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: # This is currently an "experimental" feature which provides 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: # no guarantees of backwards compatibility. It may be subject 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: # to change or removal in any patch version. Anyone opting in 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: # to this feature should strictly pin botocore. 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: host = urlparse(request.url).hostname 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: conn.proxy_headers['host'] = host 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: request_target = self._get_request_target(request.url, proxy_url) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:> urllib_response = conn.urlopen( 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: method=request.method, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: url=request_target, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: body=request.body, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: headers=request.headers, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: retries=Retry(False), 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: assert_same_host=False, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: preload_content=False, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: decode_content=False, 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: chunked=self._chunked(request.headers), 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/httpsession.py:477: 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py:802: in urlopen 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: retries = retries.increment( 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/util/retry.py:527: in increment 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: raise six.reraise(type(error), error, _stacktrace) 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/packages/six.py:770: in reraise 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: raise value 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py:716: in urlopen 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: httplib_response = self._make_request( 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py:416: in _make_request 2026-04-01T14:17:13.878 INFO:teuthology.orchestra.run.vm06.stdout: conn.request(method, url, **httplib_request_kw) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py:96: in request 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: rval = super().request(method, url, body, headers, *args, **kwargs) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:244: in request 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: super(HTTPConnection, self).request(method, url, body=body, headers=headers) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:/usr/lib64/python3.9/http/client.py:1285: in request 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self._send_request(method, url, body, headers, encode_chunked) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:/usr/lib64/python3.9/http/client.py:1331: in _send_request 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self.endheaders(body, encode_chunked=encode_chunked) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:/usr/lib64/python3.9/http/client.py:1280: in endheaders 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self._send_output(message_body, encode_chunked=encode_chunked) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py:123: in _send_output 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self.send(msg) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py:223: in send 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: return super().send(str) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:/usr/lib64/python3.9/http/client.py:980: in send 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self.connect() 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:205: in connect 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: conn = self._new_conn() 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:self = 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: def _new_conn(self): 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: """Establish a socket connection and set nodelay settings on it. 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: :return: New socket connection. 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: """ 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: extra_kw = {} 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: if self.source_address: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: extra_kw["source_address"] = self.source_address 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: if self.socket_options: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: extra_kw["socket_options"] = self.socket_options 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: conn = connection.create_connection( 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: (self._dns_host, self.port), self.timeout, **extra_kw 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: except SocketTimeout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: raise ConnectTimeoutError( 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self, 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: "Connection to %s timed out. (connect timeout=%s)" 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: % (self.host, self.timeout), 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: except SocketError as e: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:> raise NewConnectionError( 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: self, "Failed to establish a new connection: %s" % e 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:E urllib3.exceptions.NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:186: NewConnectionError 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout:During handling of the above exception, another exception occurred: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: @pytest.mark.basic_test 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: def test_dedup_dry_large_scale(): 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: #return 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: prepare_test() 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: max_copies_count=3 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: num_threads=64 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: num_files=32*1024 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: size=1*KB 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: files=[] 2026-04-01T14:17:13.879 INFO:teuthology.orchestra.run.vm06.stdout: config=TransferConfig(multipart_threshold=size, multipart_chunksize=1*MB) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: log.debug("test_dedup_dry_large_scale_new: connect to AWS ...") 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: gen_files_fixed_size(files, num_files, size, max_copies_count) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: conns=get_connections(num_threads) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: bucket_names=get_buckets(num_threads) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: for i in range(num_threads): 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: conns[i].create_bucket(Bucket=bucket_names[i]) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: threads_simple_dedup_with_tenants(files, conns, bucket_names, config, True) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: except: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: log.warning("test_dedup_dry_large_scale: failed!!") 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: finally: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: # cleanup must be executed even after a failure 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:> cleanup_all_buckets(bucket_names, conns) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py:2751: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py:496: in cleanup_all_buckets 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: delete_bucket_with_all_objects(bucket_name, conn) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py:452: in delete_bucket_with_all_objects 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: listing=conn.list_objects(Bucket=bucket_name, Marker=marker, MaxKeys=max_keys) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/client.py:602: in _api_call 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._make_api_call(operation_name, kwargs) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/context.py:123: in wrapper 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return func(*args, **kwargs) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/client.py:1060: in _make_api_call 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: http, parsed_response = self._make_request( 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/client.py:1084: in _make_request 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._endpoint.make_request(operation_model, request_dict) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:119: in make_request 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._send_request(request_dict, operation_model) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:200: in _send_request 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: while self._needs_retry( 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:360: in _needs_retry 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: responses = self._event_emitter.emit( 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/hooks.py:412: in emit 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._emitter.emit(aliased_event_name, **kwargs) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/hooks.py:256: in emit 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._emit(event_name, kwargs) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/hooks.py:239: in _emit 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: response = handler(**kwargs) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:207: in __call__ 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: if self._checker(**checker_kwargs): 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:284: in __call__ 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: should_retry = self._should_retry( 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:320: in _should_retry 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._checker(attempt_number, response, caught_exception) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:363: in __call__ 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: checker_response = checker( 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:247: in __call__ 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self._check_caught_exception( 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:416: in _check_caught_exception 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: raise caught_exception 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:279: in _do_get_response 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: http_response = self._send(request) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:383: in _send 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: return self.http_session.send(request) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:self = 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout:request = 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: def send(self, request): 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: try: 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: proxy_url = self._proxy_config.proxy_url_for(request.url) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: manager = self._get_connection_manager(request.url, proxy_url) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: conn = manager.connection_from_url(request.url) 2026-04-01T14:17:13.880 INFO:teuthology.orchestra.run.vm06.stdout: self._setup_ssl_cert(conn, request.url, self._verify) 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: if ensure_boolean( 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: os.environ.get('BOTO_EXPERIMENTAL__ADD_PROXY_HOST_HEADER', '') 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: ): 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # This is currently an "experimental" feature which provides 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # no guarantees of backwards compatibility. It may be subject 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # to change or removal in any patch version. Anyone opting in 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # to this feature should strictly pin botocore. 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: host = urlparse(request.url).hostname 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: conn.proxy_headers['host'] = host 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: request_target = self._get_request_target(request.url, proxy_url) 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: urllib_response = conn.urlopen( 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: method=request.method, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: url=request_target, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: body=request.body, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: headers=request.headers, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: retries=Retry(False), 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: assert_same_host=False, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: preload_content=False, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: decode_content=False, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: chunked=self._chunked(request.headers), 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: http_response = botocore.awsrequest.AWSResponse( 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: request.url, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: urllib_response.status, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: urllib_response.headers, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: urllib_response, 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: ) 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: if not request.stream_output: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # Cause the raw stream to be exhausted immediately. We do it 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # this way instead of using preload_content because 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: # preload_content will never buffer chunked responses 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: http_response.content 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: return http_response 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: except URLLib3SSLError as e: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: raise SSLError(endpoint_url=request.url, error=e) 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: except (NewConnectionError, socket.gaierror) as e: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout:> raise EndpointConnectionError(endpoint_url=request.url, error=e) 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout:E botocore.exceptions.EndpointConnectionError: Could not connect to the endpoint URL: "http://vm06.local:80/bnfftppgqyjpqpht-86?marker=&max-keys=1000&encoding-type=url" 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout:.tox/py/lib/python3.9/site-packages/botocore/httpsession.py:506: EndpointConnectionError 2026-04-01T14:17:13.881 INFO:teuthology.orchestra.run.vm06.stdout:----------------------------- Captured stderr call ----------------------------- 2026-04-01T14:17:14.138 INFO:teuthology.orchestra.run.vm06.stdout:ignoring --setuser ceph since I am not root 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:ignoring --setgroup ceph since I am not root 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:ignoring --setuser ceph since I am not root 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:ignoring --setgroup ceph since I am not root 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:Process Process-104: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:Traceback (most recent call last): 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 174, in _new_conn 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: conn = connection.create_connection( 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py", line 95, in create_connection 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: raise err 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py", line 85, in create_connection 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: sock.connect(sa) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:ConnectionRefusedError: [Errno 111] Connection refused 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:During handling of the above exception, another exception occurred: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:Traceback (most recent call last): 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/httpsession.py", line 477, in send 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: urllib_response = conn.urlopen( 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py", line 802, in urlopen 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: retries = retries.increment( 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/util/retry.py", line 527, in increment 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: raise six.reraise(type(error), error, _stacktrace) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/packages/six.py", line 770, in reraise 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: raise value 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py", line 716, in urlopen 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: httplib_response = self._make_request( 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py", line 416, in _make_request 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: conn.request(method, url, **httplib_request_kw) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py", line 96, in request 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: rval = super().request(method, url, body, headers, *args, **kwargs) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 244, in request 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: super(HTTPConnection, self).request(method, url, body=body, headers=headers) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/usr/lib64/python3.9/http/client.py", line 1285, in request 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: self._send_request(method, url, body, headers, encode_chunked) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/usr/lib64/python3.9/http/client.py", line 1331, in _send_request 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: self.endheaders(body, encode_chunked=encode_chunked) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/usr/lib64/python3.9/http/client.py", line 1280, in endheaders 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: self._send_output(message_body, encode_chunked=encode_chunked) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py", line 123, in _send_output 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: self.send(msg) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py", line 223, in send 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: return super().send(str) 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/usr/lib64/python3.9/http/client.py", line 980, in send 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: self.connect() 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 205, in connect 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: conn = self._new_conn() 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 186, in _new_conn 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: raise NewConnectionError( 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:urllib3.exceptions.NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:During handling of the above exception, another exception occurred: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout:Traceback (most recent call last): 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: File "/usr/lib64/python3.9/multiprocessing/process.py", line 315, in _bootstrap 2026-04-01T14:17:14.139 INFO:teuthology.orchestra.run.vm06.stdout: self.run() 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout: File "/usr/lib64/python3.9/multiprocessing/process.py", line 108, in run 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout: self._target(*self._args, **self._kwargs) 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/test_dedup.p 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:------------------------------ Captured log call ------------------------------- 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:WARNING dedup.test_dedup:test_dedup.py:2748 test_dedup_dry_large_scale: failed!! 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:=============================== warnings summary =============================== 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_with_tenants 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_multipart 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_large_mix 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_basic_with_tenants 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_multipart_with_tenants 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_small_multipart_with_tenants 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_large_scale_with_tenants 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:test_dedup.py::test_dedup_dry_large_scale 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout: /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/boto3/compat.py:89: PythonDeprecationWarning: Boto3 will no longer support Python 3.9 starting April 29, 2026. To continue receiving service updates, bug fixes, and security updates please upgrade to Python 3.10 or later. More information can be found here: https://aws.amazon.com/blogs/developer/python-support-policy-updates-for-aws-sdks-and-tools/ 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout: warnings.warn(warning, PythonDeprecationWarning) 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:=========================== short test summary info ============================ 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:FAILED test_dedup.py::test_dedup_dry_large_scale - botocore.exceptions.Endpoi... 2026-04-01T14:17:14.140 INFO:teuthology.orchestra.run.vm06.stdout:============ 1 failed, 33 passed, 8 warnings in 1289.96s (0:21:29) ============= 2026-04-01T14:17:14.335 INFO:teuthology.orchestra.run.vm06.stdout:py: exit 1 (1290.51 seconds) /home/ubuntu/cephtest/ceph/src/test/rgw/dedup> pytest -v -m 'basic_test or request_test or example_test' pid=72332 2026-04-01T14:17:14.335 INFO:teuthology.orchestra.run.vm06.stdout: py: FAIL code 1 (1293.25=setup[2.74]+cmd[1290.51] seconds) 2026-04-01T14:17:14.335 INFO:teuthology.orchestra.run.vm06.stdout: evaluation failed :( (1293.26 seconds) 2026-04-01T14:17:14.360 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:17:14.360 ERROR:teuthology.contextutil:Saw exception from nested tasks Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 30, in nested vars.append(enter()) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 191, in run_tests toxvenv_sh(ctx, remote, args, label="dedup tests against rgw") File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 165, in toxvenv_sh return remote.sh(['source', activate, run.Raw('&&')] + args, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 97, in sh proc = self.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed (dedup tests against rgw) on vm06 with status 1: "source /home/ubuntu/cephtest/tox-venv/bin/activate && cd /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/ && DEDUPTESTS_CONF=./deduptests.client.0.conf tox -- -v -m 'basic_test or request_test or example_test'" 2026-04-01T14:17:14.362 INFO:tasks.dedup_tests:Removing dedup-tests.conf file... 2026-04-01T14:17:14.362 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/deduptests.client.0.conf 2026-04-01T14:17:14.383 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user rm --uid foo.client.0 --purge-data --cluster ceph 2026-04-01T14:17:14.458 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T14:17:14.459 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T14:22:14.460 INFO:teuthology.orchestra.run.vm06.stderr:failed to fetch mon config (--no-mon-config to skip) 2026-04-01T14:22:14.462 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:22:14.462 INFO:tasks.dedup_tests:Removing dedup-tests... 2026-04-01T14:22:14.462 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/ceph 2026-04-01T14:22:14.996 ERROR:teuthology.run_tasks:Saw exception from tasks. Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 107, in create_users yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 30, in nested vars.append(enter()) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 191, in run_tests toxvenv_sh(ctx, remote, args, label="dedup tests against rgw") File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 165, in toxvenv_sh return remote.sh(['source', activate, run.Raw('&&')] + args, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 97, in sh proc = self.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed (dedup tests against rgw) on vm06 with status 1: "source /home/ubuntu/cephtest/tox-venv/bin/activate && cd /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/ && DEDUPTESTS_CONF=./deduptests.client.0.conf tox -- -v -m 'basic_test or request_test or example_test'" During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 112, in run_tasks manager.__enter__() File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 240, in task with contextutil.nested( File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 45, in download yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/dedup_tests.py", line 114, in create_users ctx.cluster.only(client).run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/cluster.py", line 85, in run procs = [remote.run(**kwargs, wait=_wait) for remote in remotes] File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/cluster.py", line 85, in procs = [remote.run(**kwargs, wait=_wait) for remote in remotes] File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user rm --uid foo.client.0 --purge-data --cluster ceph' 2026-04-01T14:22:14.997 DEBUG:teuthology.run_tasks:Unwinding manager dedup-tests 2026-04-01T14:22:15.003 DEBUG:teuthology.run_tasks:Unwinding manager tox 2026-04-01T14:22:15.006 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/tox-venv 2026-04-01T14:22:15.076 DEBUG:teuthology.run_tasks:Unwinding manager tox 2026-04-01T14:22:15.078 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/tox-venv 2026-04-01T14:22:15.093 DEBUG:teuthology.run_tasks:Unwinding manager rgw 2026-04-01T14:22:15.096 DEBUG:tasks.rgw.client.0:waiting for process to exit 2026-04-01T14:22:15.096 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:22:15.097 INFO:tasks.rgw.client.0:Stopped 2026-04-01T14:22:15.097 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock 2026-04-01T14:22:15.149 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/vault-root-token 2026-04-01T14:22:15.222 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /home/ubuntu/cephtest/url_file 2026-04-01T14:22:15.286 INFO:tasks.util.rgw:rgwadmin: client.0 : ['gc', 'process', '--include-all'] 2026-04-01T14:22:15.286 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'gc', 'process', '--include-all'] 2026-04-01T14:22:15.286 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all 2026-04-01T14:22:15.360 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-01T14:22:15.360 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-01T14:27:15.362 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-01T14:27:15.361+0000 7fac83f63900 0 monclient(hunting): authenticate timed out after 300 2026-04-01T14:27:15.362 INFO:teuthology.orchestra.run.vm06.stderr:failed to fetch mon config (--no-mon-config to skip) 2026-04-01T14:27:15.363 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:27:15.364 ERROR:teuthology.run_tasks:Manager failed: rgw Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' 2026-04-01T14:27:15.364 DEBUG:teuthology.run_tasks:Unwinding manager openssl_keys 2026-04-01T14:27:15.366 DEBUG:teuthology.run_tasks:Unwinding manager ceph 2026-04-01T14:27:15.368 INFO:tasks.ceph.ceph_manager.ceph:waiting for clean 2026-04-01T14:27:15.369 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json 2026-04-01T14:29:15.435 DEBUG:teuthology.orchestra.run:got remote process result: 124 2026-04-01T14:29:15.436 ERROR:teuthology.contextutil:Saw exception from nested tasks Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' 2026-04-01T14:29:15.436 INFO:teuthology.misc:Shutting down mds daemons... 2026-04-01T14:29:15.436 INFO:teuthology.misc:Shutting down osd daemons... 2026-04-01T14:29:15.436 DEBUG:tasks.ceph.osd.0:waiting for process to exit 2026-04-01T14:29:15.436 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.0:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.1:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.1:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.2:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.2:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.3:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.3:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.4:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.4:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.5:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.5:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.6:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.6:Stopped 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.osd.7:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 INFO:tasks.ceph.osd.7:Stopped 2026-04-01T14:29:15.437 INFO:teuthology.misc:Shutting down mgr daemons... 2026-04-01T14:29:15.437 DEBUG:tasks.ceph.mgr.y:waiting for process to exit 2026-04-01T14:29:15.437 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.437 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:29:15.437 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster ceph -i y' 2026-04-01T14:29:15.438 INFO:tasks.ceph.mgr.y:Stopped 2026-04-01T14:29:15.438 DEBUG:tasks.ceph.mgr.x:waiting for process to exit 2026-04-01T14:29:15.438 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.438 INFO:tasks.ceph.mgr.x:Stopped 2026-04-01T14:29:15.438 INFO:teuthology.misc:Shutting down mon daemons... 2026-04-01T14:29:15.438 DEBUG:tasks.ceph.mon.a:waiting for process to exit 2026-04-01T14:29:15.438 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.438 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:29:15.438 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i a' 2026-04-01T14:29:15.438 INFO:tasks.ceph.mon.a:Stopped 2026-04-01T14:29:15.438 DEBUG:tasks.ceph.mon.c:waiting for process to exit 2026-04-01T14:29:15.438 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.438 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:29:15.438 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i c' 2026-04-01T14:29:15.438 INFO:tasks.ceph.mon.c:Stopped 2026-04-01T14:29:15.438 DEBUG:tasks.ceph.mon.b:waiting for process to exit 2026-04-01T14:29:15.438 INFO:teuthology.orchestra.run:waiting for 300 2026-04-01T14:29:15.438 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:29:15.438 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm08 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i b' 2026-04-01T14:29:15.439 INFO:tasks.ceph.mon.b:Stopped 2026-04-01T14:29:15.439 INFO:tasks.ceph:Checking cluster log for badness... 2026-04-01T14:29:15.439 DEBUG:teuthology.orchestra.run.vm06:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v '\(PG_AVAILABILITY\)' | egrep -v '\(PG_DEGRADED\)' | egrep -v '\(POOL_APP_NOT_ENABLED\)' | egrep -v 'not have an application enabled' | head -n 1 2026-04-01T14:29:15.466 INFO:teuthology.orchestra.run.vm06.stdout:2026-04-01T14:05:28.133579+0000 mon.a (mon.0) 679 : cluster [ERR] Health check failed: mon c is very low on available space (MON_DISK_CRIT) 2026-04-01T14:29:15.467 WARNING:tasks.ceph:Found errors (ERR|WRN|SEC) in cluster log 2026-04-01T14:29:15.467 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-0 on ubuntu@vm06.local 2026-04-01T14:29:15.467 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-0 2026-04-01T14:29:15.591 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-1 on ubuntu@vm06.local 2026-04-01T14:29:15.591 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-1 2026-04-01T14:29:15.680 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-2 on ubuntu@vm06.local 2026-04-01T14:29:15.680 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-2 2026-04-01T14:29:15.765 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-3 on ubuntu@vm06.local 2026-04-01T14:29:15.765 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-3 2026-04-01T14:29:15.861 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-4 on ubuntu@vm08.local 2026-04-01T14:29:15.861 DEBUG:teuthology.orchestra.run.vm08:> sync && sudo umount -f /var/lib/ceph/osd/ceph-4 2026-04-01T14:29:15.984 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-5 on ubuntu@vm08.local 2026-04-01T14:29:15.984 DEBUG:teuthology.orchestra.run.vm08:> sync && sudo umount -f /var/lib/ceph/osd/ceph-5 2026-04-01T14:29:16.082 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-6 on ubuntu@vm08.local 2026-04-01T14:29:16.082 DEBUG:teuthology.orchestra.run.vm08:> sync && sudo umount -f /var/lib/ceph/osd/ceph-6 2026-04-01T14:29:16.189 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-7 on ubuntu@vm08.local 2026-04-01T14:29:16.189 DEBUG:teuthology.orchestra.run.vm08:> sync && sudo umount -f /var/lib/ceph/osd/ceph-7 2026-04-01T14:29:16.284 INFO:tasks.ceph:Archiving mon data... 2026-04-01T14:29:16.285 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/mon/ceph-a to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/data/mon.a.tgz 2026-04-01T14:29:16.285 DEBUG:teuthology.orchestra.run.vm06:> mktemp 2026-04-01T14:29:16.302 INFO:teuthology.orchestra.run.vm06.stdout:/tmp/tmp.nsagj099Gw 2026-04-01T14:29:16.302 DEBUG:teuthology.orchestra.run.vm06:> sudo tar cz -f - -C /var/lib/ceph/mon/ceph-a -- . > /tmp/tmp.nsagj099Gw 2026-04-01T14:29:16.444 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0666 /tmp/tmp.nsagj099Gw 2026-04-01T14:29:16.525 DEBUG:teuthology.orchestra.remote:vm06:/tmp/tmp.nsagj099Gw is 504KB 2026-04-01T14:29:16.586 DEBUG:teuthology.orchestra.run.vm06:> rm -fr /tmp/tmp.nsagj099Gw 2026-04-01T14:29:16.601 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/mon/ceph-c to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/data/mon.c.tgz 2026-04-01T14:29:16.601 DEBUG:teuthology.orchestra.run.vm06:> mktemp 2026-04-01T14:29:16.657 INFO:teuthology.orchestra.run.vm06.stdout:/tmp/tmp.9H3Xla2Thj 2026-04-01T14:29:16.657 DEBUG:teuthology.orchestra.run.vm06:> sudo tar cz -f - -C /var/lib/ceph/mon/ceph-c -- . > /tmp/tmp.9H3Xla2Thj 2026-04-01T14:29:16.798 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0666 /tmp/tmp.9H3Xla2Thj 2026-04-01T14:29:16.876 DEBUG:teuthology.orchestra.remote:vm06:/tmp/tmp.9H3Xla2Thj is 527KB 2026-04-01T14:29:16.935 DEBUG:teuthology.orchestra.run.vm06:> rm -fr /tmp/tmp.9H3Xla2Thj 2026-04-01T14:29:16.949 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/mon/ceph-b to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/data/mon.b.tgz 2026-04-01T14:29:16.949 DEBUG:teuthology.orchestra.run.vm08:> mktemp 2026-04-01T14:29:16.966 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:29:16.966 INFO:teuthology.orchestra.run.vm08.stderr:mktemp: failed to create file via template ‘/tmp/tmp.XXXXXXXXXX’: No space left on device 2026-04-01T14:29:17.005 INFO:teuthology.util.scanner:summary_data or yaml_file is empty! 2026-04-01T14:29:17.023 INFO:teuthology.util.scanner:summary_data or yaml_file is empty! 2026-04-01T14:29:17.039 INFO:teuthology.util.scanner:summary_data or yaml_file is empty! 2026-04-01T14:29:17.039 INFO:tasks.ceph:Archiving crash dumps... 2026-04-01T14:29:17.040 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/crash to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/remote/vm06/crash 2026-04-01T14:29:17.040 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/crash -- . 2026-04-01T14:29:17.070 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/crash to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/remote/vm08/crash 2026-04-01T14:29:17.070 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/lib/ceph/crash -- . 2026-04-01T14:29:17.094 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/crash to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/remote/vm09/crash 2026-04-01T14:29:17.094 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/crash -- . 2026-04-01T14:29:17.126 INFO:tasks.ceph:Compressing logs... 2026-04-01T14:29:17.126 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-04-01T14:29:17.127 DEBUG:teuthology.orchestra.run.vm08:> time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-04-01T14:29:17.137 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-04-01T14:29:17.151 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.tmp-client.admin.58211.log 2026-04-01T14:29:17.151 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.0.log 2026-04-01T14:29:17.152 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph.tmp-client.admin.58211.log: 0.0% -- replaced with /var/log/ceph/ceph.tmp-client.admin.58211.log.gz 2026-04-01T14:29:17.152 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.1.log 2026-04-01T14:29:17.152 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-osd.0.log: gzip -5 --verbose -- /var/log/ceph/ceph-osd.2.log 2026-04-01T14:29:17.157 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-osd.1.log: /var/log/ceph/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/ceph-osd.3.log 2026-04-01T14:29:17.157 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-osd.3.log: gzip -5 --verbose -- /var/log/ceph/ceph-mon.a.log 2026-04-01T14:29:17.159 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.4.log 2026-04-01T14:29:17.159 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.5.log 2026-04-01T14:29:17.159 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.6.log 2026-04-01T14:29:17.160 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.7.log 2026-04-01T14:29:17.160 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-osd.4.log.gz: No space left on device 2026-04-01T14:29:17.160 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-osd.6.log.gz: No space left on device 2026-04-01T14:29:17.160 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-osd.5.log.gz: No space left on device 2026-04-01T14:29:17.160 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mon.b.log 2026-04-01T14:29:17.160 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.log 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-osd.7.log.gz: No space left on device 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-mon.b.log.gz: No space left on device 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mgr.x.log 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.57416.log 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph.log.gz: No space left on device 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-mgr.x.log.gz: No space left on device 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.57464.log 2026-04-01T14:29:17.161 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.57416.log.gz: No space left on device 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.audit.log 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.57512.log 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.57464.log.gz: No space left on device 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:gzip: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.57560.log 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph.audit.log.gz: No space left on device 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.57512.log.gz: No space left on device 2026-04-01T14:29:17.162 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.57608.log 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.57560.log.gz: No space left on device 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.57656.log 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:gzip: gzip/var/log/ceph/ceph-client.admin.57608.log.gz: No space left on device 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr: -5 --verbose -- /var/log/ceph/ceph-client.admin.57704.log 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:gzipgzip: -5 --verbose -- /var/log/ceph/ceph-client.admin.57752.log 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.admin.57656.log.gz: No space left on device 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.57704.log.gz: No space left on device 2026-04-01T14:29:17.163 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.61387.log 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip: gzip -5 --verbose --/var/log/ceph/ceph-client.admin.57752.log.gz: No space left on device 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr: /var/log/ceph/ceph-client.admin.61435.log 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.61387.log.gz: No space left on device 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.61483.log 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.61435.log.gz: No space left on device 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.61531.log 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.admin.61483.log.gz: No space left on device 2026-04-01T14:29:17.164 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.61554.log 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:gzip: gzip/var/log/ceph/ceph-client.admin.61531.log.gz: No space left on device 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr: -5 --verbose -- /var/log/ceph/ceph-client.1.61662.log 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:gzip: gzip -5 --verbose -- /var/log/ceph/ceph-client.1.61765.log 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.1.61554.log.gz: No space left on device 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.1.61662.log.gz: No space left on device 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.61868.log 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.1.61765.log.gz: No space left on device 2026-04-01T14:29:17.165 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.61971.log 2026-04-01T14:29:17.166 INFO:teuthology.orchestra.run.vm08.stderr:gzip: gzip -5 --verbose -- /var/log/ceph/rgw.ceph.client.1.log 2026-04-01T14:29:17.166 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.1.61868.log.gz: No space left on device 2026-04-01T14:29:17.166 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ceph-client.1.61971.log.gz: No space left on device 2026-04-01T14:29:17.166 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ops-log-ceph-client.1.log 2026-04-01T14:29:17.166 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/rgw.ceph.client.1.log.gz: No space left on device 2026-04-01T14:29:17.166 INFO:teuthology.orchestra.run.vm08.stderr:gzip: /var/log/ceph/ops-log-ceph-client.1.log.gz: No space left on device 2026-04-01T14:29:17.168 INFO:teuthology.orchestra.run.vm08.stderr: 2026-04-01T14:29:17.168 INFO:teuthology.orchestra.run.vm08.stderr:real 0m0.019s 2026-04-01T14:29:17.168 INFO:teuthology.orchestra.run.vm08.stderr:user 0m0.009s 2026-04-01T14:29:17.168 INFO:teuthology.orchestra.run.vm08.stderr:sys 0m0.024s 2026-04-01T14:29:17.172 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mon.c.log 2026-04-01T14:29:17.179 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64617.log 2026-04-01T14:29:17.192 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/ceph.log 2026-04-01T14:29:17.192 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.64617.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64617.log.gz 2026-04-01T14:29:17.193 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.51083.log 2026-04-01T14:29:17.193 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.51131.log 2026-04-01T14:29:17.194 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.admin.51083.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.51083.log.gz 2026-04-01T14:29:17.194 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.51179.log 2026-04-01T14:29:17.194 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.51227.log 2026-04-01T14:29:17.194 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.admin.51131.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.51131.log.gz 2026-04-01T14:29:17.194 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.admin.51179.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.2.51250.log 2026-04-01T14:29:17.194 INFO:teuthology.orchestra.run.vm09.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.51179.log.gz 2026-04-01T14:29:17.195 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5/var/log/ceph/ceph-client.admin.51227.log: --verbose -- /var/log/ceph/ceph-client.2.51358.log 2026-04-01T14:29:17.195 INFO:teuthology.orchestra.run.vm09.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.51227.log.gz 2026-04-01T14:29:17.195 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.2.51461.log 2026-04-01T14:29:17.195 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.2.51250.log: 83.2% -- replaced with /var/log/ceph/ceph-client.2.51250.log.gz 2026-04-01T14:29:17.195 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.2.51564.log 2026-04-01T14:29:17.195 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.2.51358.log: 45.6% -- replaced with /var/log/ceph/ceph-client.2.51358.log.gz 2026-04-01T14:29:17.196 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.2.51461.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.2.51667.log 2026-04-01T14:29:17.196 INFO:teuthology.orchestra.run.vm09.stderr: 44.2% -- replaced with /var/log/ceph/ceph-client.2.51461.log.gz 2026-04-01T14:29:17.196 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.2.51564.log: gzip -5 --verbose -- /var/log/ceph/rgw.ceph.client.2.log 2026-04-01T14:29:17.196 INFO:teuthology.orchestra.run.vm09.stderr: 45.3% -- replaced with /var/log/ceph/ceph-client.2.51564.log.gz 2026-04-01T14:29:17.196 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ceph-client.2.51667.log: gzip -5 --verbose -- /var/log/ceph/ops-log-ceph-client.2.log 2026-04-01T14:29:17.196 INFO:teuthology.orchestra.run.vm09.stderr: 45.3% -- replaced with /var/log/ceph/ceph-client.2.51667.log.gz 2026-04-01T14:29:17.197 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/rgw.ceph.client.2.log: /var/log/ceph/ops-log-ceph-client.2.log: 35.3% -- replaced with /var/log/ceph/ops-log-ceph-client.2.log.gz 2026-04-01T14:29:17.207 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64709.log 2026-04-01T14:29:17.211 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph.log: 92.9% -- replaced with /var/log/ceph/ceph.log.gz 2026-04-01T14:29:17.216 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mgr.y.log 2026-04-01T14:29:17.216 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.64709.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64709.log.gz 2026-04-01T14:29:17.228 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.audit.log 2026-04-01T14:29:17.238 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-mgr.y.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64994.log 2026-04-01T14:29:17.239 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph.audit.log: 94.5% -- replaced with /var/log/ceph/ceph-mgr.y.log.gz 2026-04-01T14:29:17.240 INFO:teuthology.orchestra.run.vm06.stderr: 94.5% -- replaced with /var/log/ceph/ceph.audit.log.gz 2026-04-01T14:29:17.251 INFO:teuthology.orchestra.run.vm09.stderr: 91.2% -- replaced with /var/log/ceph/rgw.ceph.client.2.log.gz 2026-04-01T14:29:17.252 INFO:teuthology.orchestra.run.vm09.stderr: 2026-04-01T14:29:17.252 INFO:teuthology.orchestra.run.vm09.stderr:real 0m0.070s 2026-04-01T14:29:17.252 INFO:teuthology.orchestra.run.vm09.stderr:user 0m0.062s 2026-04-01T14:29:17.252 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.021s 2026-04-01T14:29:17.255 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68221.log 2026-04-01T14:29:17.255 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.64994.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64994.log.gz 2026-04-01T14:29:17.265 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68444.log 2026-04-01T14:29:17.265 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68221.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68221.log.gz 2026-04-01T14:29:17.271 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68516.log 2026-04-01T14:29:17.271 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68444.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68444.log.gz 2026-04-01T14:29:17.281 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68543.log 2026-04-01T14:29:17.281 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68516.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68516.log.gz 2026-04-01T14:29:17.287 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68634.log 2026-04-01T14:29:17.287 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68543.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68543.log.gz 2026-04-01T14:29:17.297 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68684.log 2026-04-01T14:29:17.297 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68634.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68634.log.gz 2026-04-01T14:29:17.303 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68734.log 2026-04-01T14:29:17.303 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68684.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68684.log.gz 2026-04-01T14:29:17.313 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68784.log 2026-04-01T14:29:17.313 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68734.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68734.log.gz 2026-04-01T14:29:17.319 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69025.log 2026-04-01T14:29:17.319 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68784.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68784.log.gz 2026-04-01T14:29:17.329 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69024.log 2026-04-01T14:29:17.329 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69025.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69025.log.gz 2026-04-01T14:29:17.335 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69027.log 2026-04-01T14:29:17.335 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69024.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69024.log.gz 2026-04-01T14:29:17.344 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69009.log 2026-04-01T14:29:17.344 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69026.log 2026-04-01T14:29:17.345 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69027.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69027.log.gz 2026-04-01T14:29:17.345 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69023.log 2026-04-01T14:29:17.345 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69009.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69030.log 2026-04-01T14:29:17.345 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69023.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69023.log.gz 2026-04-01T14:29:17.351 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68998.log 2026-04-01T14:29:17.357 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69030.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69009.log.gz 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69030.log.gz 2026-04-01T14:29:17.357 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-01T14:29:17.360 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69306.log 2026-04-01T14:29:17.360 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.68998.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68998.log.gz 2026-04-01T14:29:17.366 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69026.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69026.log.gz 2026-04-01T14:29:17.376 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69399.log 2026-04-01T14:29:17.376 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69306.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69306.log.gz 2026-04-01T14:29:17.381 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69381.log 2026-04-01T14:29:17.383 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69399.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69399.log.gz 2026-04-01T14:29:17.394 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69394.log 2026-04-01T14:29:17.394 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69381.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69381.log.gz 2026-04-01T14:29:17.400 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69380.log 2026-04-01T14:29:17.400 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69394.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69394.log.gz 2026-04-01T14:29:17.414 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69398.log 2026-04-01T14:29:17.414 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69380.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69380.log.gz 2026-04-01T14:29:17.420 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69406.log 2026-04-01T14:29:17.420 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69398.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69398.log.gz 2026-04-01T14:29:17.430 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69452.log 2026-04-01T14:29:17.430 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69406.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69406.log.gz 2026-04-01T14:29:17.437 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69634.log 2026-04-01T14:29:17.437 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69452.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69452.log.gz 2026-04-01T14:29:17.447 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69682.log 2026-04-01T14:29:17.447 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69634.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69634.log.gz 2026-04-01T14:29:17.452 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69711.log 2026-04-01T14:29:17.452 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69682.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69682.log.gz 2026-04-01T14:29:17.463 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69862.log 2026-04-01T14:29:17.463 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69711.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69711.log.gz 2026-04-01T14:29:17.469 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69866.log 2026-04-01T14:29:17.469 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69862.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69862.log.gz 2026-04-01T14:29:17.476 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69863.log 2026-04-01T14:29:17.476 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69866.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69866.log.gz 2026-04-01T14:29:17.487 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69855.log 2026-04-01T14:29:17.487 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69863.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69863.log.gz 2026-04-01T14:29:17.494 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69871.log 2026-04-01T14:29:17.494 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69855.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69855.log.gz 2026-04-01T14:29:17.504 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70034.log 2026-04-01T14:29:17.504 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.69871.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69871.log.gz 2026-04-01T14:29:17.509 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70082.log 2026-04-01T14:29:17.509 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70034.log: 92.4% -- replaced with /var/log/ceph/ceph-mon.c.log.gz 2026-04-01T14:29:17.511 INFO:teuthology.orchestra.run.vm06.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70034.log.gz 2026-04-01T14:29:17.524 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70158.log 2026-04-01T14:29:17.524 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70082.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70082.log.gz 2026-04-01T14:29:17.538 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70157.log 2026-04-01T14:29:17.538 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70158.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70158.log.gz 2026-04-01T14:29:17.543 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70216.log 2026-04-01T14:29:17.550 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70157.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70213.log 2026-04-01T14:29:17.550 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70216.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70216.log.gz 2026-04-01T14:29:17.552 INFO:teuthology.orchestra.run.vm06.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70157.log.gz 2026-04-01T14:29:17.568 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70354.log 2026-04-01T14:29:17.568 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70213.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70213.log.gz 2026-04-01T14:29:17.576 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70352.log 2026-04-01T14:29:17.576 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70354.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70354.log.gz 2026-04-01T14:29:17.586 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70434.log 2026-04-01T14:29:17.586 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70352.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70352.log.gz 2026-04-01T14:29:17.594 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70484.log 2026-04-01T14:29:17.594 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70434.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70434.log.gz 2026-04-01T14:29:17.601 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70534.log 2026-04-01T14:29:17.601 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70484.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70484.log.gz 2026-04-01T14:29:17.609 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70582.log 2026-04-01T14:29:17.609 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70534.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70534.log.gz 2026-04-01T14:29:17.616 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70632.log 2026-04-01T14:29:17.616 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70582.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70582.log.gz 2026-04-01T14:29:17.623 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70680.log 2026-04-01T14:29:17.623 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70632.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70632.log.gz 2026-04-01T14:29:17.637 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70730.log 2026-04-01T14:29:17.637 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70680.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70680.log.gz 2026-04-01T14:29:17.639 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.70753.log 2026-04-01T14:29:17.639 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.70730.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70730.log.gz 2026-04-01T14:29:17.640 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.70869.log 2026-04-01T14:29:17.640 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.70753.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.0.70972.log 2026-04-01T14:29:17.640 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.70869.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.0.71075.log 2026-04-01T14:29:17.640 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.70972.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.0.71178.log 2026-04-01T14:29:17.641 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.71075.log: gzip -5 --verbose -- /var/log/ceph/rgw.ceph.client.0.log 2026-04-01T14:29:17.641 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.71178.log: gzip -5 --verbose -- /var/log/ceph/ops-log-ceph-client.0.log 2026-04-01T14:29:17.642 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/rgw.ceph.client.0.log: 45.3% -- replaced with /var/log/ceph/ceph-client.0.70972.log.gz 2026-04-01T14:29:17.642 INFO:teuthology.orchestra.run.vm06.stderr: 45.3% -- replaced with /var/log/ceph/ceph-client.0.71178.log.gz 2026-04-01T14:29:17.643 INFO:teuthology.orchestra.run.vm06.stderr: 95.1% -- replaced with /var/log/ceph/ceph-client.0.70753.log.gz 2026-04-01T14:29:17.646 INFO:teuthology.orchestra.run.vm06.stderr: 45.3% -- replaced with /var/log/ceph/ceph-client.0.70869.log.gz 2026-04-01T14:29:17.646 INFO:teuthology.orchestra.run.vm06.stderr: 45.6% -- replaced with /var/log/ceph/ceph-client.0.71075.log.gz 2026-04-01T14:29:17.652 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.72158.log 2026-04-01T14:29:17.662 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ops-log-ceph-client.0.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72350.log 2026-04-01T14:29:17.662 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.72158.log: 84.6% -- replaced with /var/log/ceph/ceph-client.0.72158.log.gz 2026-04-01T14:29:17.673 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72491.log 2026-04-01T14:29:17.681 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72350.log: 83.3% -- replaced with /var/log/ceph/ceph-client.admin.72350.log.gz 2026-04-01T14:29:17.691 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72525.log 2026-04-01T14:29:17.691 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72491.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72491.log.gz 2026-04-01T14:29:17.704 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72559.log 2026-04-01T14:29:17.704 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72525.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72525.log.gz 2026-04-01T14:29:17.726 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72656.log 2026-04-01T14:29:17.726 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72559.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.72559.log.gz 2026-04-01T14:29:17.739 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72753.log 2026-04-01T14:29:17.745 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72656.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72916.log 2026-04-01T14:29:17.745 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72753.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.72656.log.gz 2026-04-01T14:29:17.750 INFO:teuthology.orchestra.run.vm06.stderr: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.72753.log.gz 2026-04-01T14:29:17.761 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72950.log 2026-04-01T14:29:17.761 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72916.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72916.log.gz 2026-04-01T14:29:17.771 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72984.log 2026-04-01T14:29:17.771 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72950.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72950.log.gz 2026-04-01T14:29:17.791 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73095.log 2026-04-01T14:29:17.801 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.72984.log: 82.6% -- replaced with /var/log/ceph/ceph-client.admin.72984.log.gz 2026-04-01T14:29:17.811 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73193.log 2026-04-01T14:29:17.811 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.73095.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.73095.log.gz 2026-04-01T14:29:17.830 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73290.log 2026-04-01T14:29:17.830 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.73193.log: 93.2% -- replaced with /var/log/ceph/ceph-client.admin.73193.log.gz 2026-04-01T14:29:17.848 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73324.log 2026-04-01T14:29:17.848 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.73290.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73290.log.gz 2026-04-01T14:29:17.862 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73360.log 2026-04-01T14:29:17.862 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.73324.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73324.log.gz 2026-04-01T14:29:17.876 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73394.log 2026-04-01T14:29:17.876 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.73360.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73360.log.gz 2026-04-01T14:29:17.890 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74268.log 2026-04-01T14:29:17.890 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.73394.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73394.log.gz 2026-04-01T14:29:17.900 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74302.log 2026-04-01T14:29:17.900 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74268.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74268.log.gz 2026-04-01T14:29:17.910 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74336.log 2026-04-01T14:29:17.910 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74302.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74302.log.gz 2026-04-01T14:29:17.923 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74439.log 2026-04-01T14:29:17.924 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74336.log: 82.9% -- replaced with /var/log/ceph/ceph-client.admin.74336.log.gz 2026-04-01T14:29:17.937 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74537.log 2026-04-01T14:29:17.947 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74439.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.74439.log.gz 2026-04-01T14:29:17.957 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74634.log 2026-04-01T14:29:17.962 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74537.log: 96.7% -- replaced with /var/log/ceph/ceph-client.admin.74537.log.gz 2026-04-01T14:29:17.973 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74668.log 2026-04-01T14:29:17.973 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74634.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74634.log.gz 2026-04-01T14:29:17.987 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74703.log 2026-04-01T14:29:17.987 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74668.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74668.log.gz 2026-04-01T14:29:17.995 INFO:teuthology.orchestra.run.vm06.stderr: 91.3% -- replaced with /var/log/ceph/ceph-mon.a.log.gz 2026-04-01T14:29:18.001 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74737.log 2026-04-01T14:29:18.001 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74703.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74703.log.gz 2026-04-01T14:29:18.015 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74863.log 2026-04-01T14:29:18.015 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74737.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74737.log.gz 2026-04-01T14:29:18.029 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74897.log 2026-04-01T14:29:18.029 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74863.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74863.log.gz 2026-04-01T14:29:18.044 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74931.log 2026-04-01T14:29:18.044 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74897.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74897.log.gz 2026-04-01T14:29:18.058 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75034.log 2026-04-01T14:29:18.058 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.74931.log: 83.0% -- replaced with /var/log/ceph/ceph-client.admin.74931.log.gz 2026-04-01T14:29:18.068 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75132.log 2026-04-01T14:29:18.068 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75034.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.75034.log.gz 2026-04-01T14:29:18.073 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75229.log 2026-04-01T14:29:18.083 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75132.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75263.log 2026-04-01T14:29:18.083 INFO:teuthology.orchestra.run.vm06.stderr: 93.0% -- replaced with /var/log/ceph/ceph-client.admin.75132.log.gz 2026-04-01T14:29:18.083 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75229.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75229.log.gz 2026-04-01T14:29:18.101 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75299.log 2026-04-01T14:29:18.101 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75263.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75263.log.gz 2026-04-01T14:29:18.111 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75333.log 2026-04-01T14:29:18.111 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75299.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75299.log.gz 2026-04-01T14:29:18.116 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75559.log 2026-04-01T14:29:18.125 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75333.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75333.log.gz 2026-04-01T14:29:18.126 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75593.log 2026-04-01T14:29:18.126 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75559.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75559.log.gz 2026-04-01T14:29:18.140 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75627.log 2026-04-01T14:29:18.140 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75593.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75593.log.gz 2026-04-01T14:29:18.150 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75730.log 2026-04-01T14:29:18.150 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75627.log: 82.7% -- replaced with /var/log/ceph/ceph-client.admin.75627.log.gz 2026-04-01T14:29:18.155 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75828.log 2026-04-01T14:29:18.165 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75730.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.75730.log.gz 2026-04-01T14:29:18.165 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75925.log 2026-04-01T14:29:18.165 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75828.log: 89.4% -- replaced with /var/log/ceph/ceph-client.admin.75828.log.gz 2026-04-01T14:29:18.170 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75959.log 2026-04-01T14:29:18.179 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75925.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75925.log.gz 2026-04-01T14:29:18.180 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75994.log 2026-04-01T14:29:18.180 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75959.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75959.log.gz 2026-04-01T14:29:18.194 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76028.log 2026-04-01T14:29:18.194 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.75994.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75994.log.gz 2026-04-01T14:29:18.204 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76140.log 2026-04-01T14:29:18.204 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76028.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76028.log.gz 2026-04-01T14:29:18.209 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76174.log 2026-04-01T14:29:18.218 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76140.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76140.log.gz 2026-04-01T14:29:18.219 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76208.log 2026-04-01T14:29:18.219 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76174.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76174.log.gz 2026-04-01T14:29:18.225 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76311.log 2026-04-01T14:29:18.233 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76208.log: 83.0% -- replaced with /var/log/ceph/ceph-client.admin.76208.log.gz 2026-04-01T14:29:18.234 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76409.log 2026-04-01T14:29:18.235 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76311.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.76311.log.gz 2026-04-01T14:29:18.248 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76506.log 2026-04-01T14:29:18.248 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76409.log: 93.2% -- replaced with /var/log/ceph/ceph-client.admin.76409.log.gz 2026-04-01T14:29:18.258 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76540.log 2026-04-01T14:29:18.258 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76506.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76506.log.gz 2026-04-01T14:29:18.264 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76576.log 2026-04-01T14:29:18.264 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76540.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76540.log.gz 2026-04-01T14:29:18.274 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76610.log 2026-04-01T14:29:18.274 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76576.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76576.log.gz 2026-04-01T14:29:18.279 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77412.log 2026-04-01T14:29:18.288 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.76610.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76610.log.gz 2026-04-01T14:29:18.289 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77446.log 2026-04-01T14:29:18.289 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77412.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.77412.log.gz 2026-04-01T14:29:18.296 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77480.log 2026-04-01T14:29:18.296 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77446.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.77446.log.gz 2026-04-01T14:29:18.312 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77585.log 2026-04-01T14:29:18.312 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77480.log: 82.8% -- replaced with /var/log/ceph/ceph-client.admin.77480.log.gz 2026-04-01T14:29:18.317 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77683.log 2026-04-01T14:29:18.327 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77585.log: 84.8% -- replaced with /var/log/ceph/ceph-client.admin.77585.log.gz 2026-04-01T14:29:18.327 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77780.log 2026-04-01T14:29:18.328 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77683.log: 90.3% -- replaced with /var/log/ceph/ceph-client.admin.77683.log.gz 2026-04-01T14:29:18.341 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77814.log 2026-04-01T14:29:18.341 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77780.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.77780.log.gz 2026-04-01T14:29:18.351 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77850.log 2026-04-01T14:29:18.351 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77814.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.77814.log.gz 2026-04-01T14:29:18.356 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77884.log 2026-04-01T14:29:18.365 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77850.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.77850.log.gz 2026-04-01T14:29:18.366 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77918.log 2026-04-01T14:29:18.366 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77884.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.77884.log.gz 2026-04-01T14:29:18.380 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78015.log 2026-04-01T14:29:18.380 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.77918.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.77918.log.gz 2026-04-01T14:29:18.390 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78112.log 2026-04-01T14:29:18.396 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78015.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78461.log 2026-04-01T14:29:18.396 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78112.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78015.log.gz 2026-04-01T14:29:18.406 INFO:teuthology.orchestra.run.vm06.stderr: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78112.log.gz 2026-04-01T14:29:18.416 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78495.log 2026-04-01T14:29:18.416 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78461.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.78461.log.gz 2026-04-01T14:29:18.421 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78529.log 2026-04-01T14:29:18.430 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78495.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.78495.log.gz 2026-04-01T14:29:18.431 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78632.log 2026-04-01T14:29:18.431 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78529.log: 83.0% -- replaced with /var/log/ceph/ceph-client.admin.78529.log.gz 2026-04-01T14:29:18.436 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78730.log 2026-04-01T14:29:18.444 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78632.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78827.log 2026-04-01T14:29:18.445 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78730.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.78632.log.gz 2026-04-01T14:29:18.446 INFO:teuthology.orchestra.run.vm06.stderr: 94.1% -- replaced with /var/log/ceph/ceph-client.admin.78730.log.gz 2026-04-01T14:29:18.461 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78861.log 2026-04-01T14:29:18.461 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78827.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.78827.log.gz 2026-04-01T14:29:18.471 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78897.log 2026-04-01T14:29:18.471 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78861.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.78861.log.gz 2026-04-01T14:29:18.476 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78931.log 2026-04-01T14:29:18.485 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78897.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.78897.log.gz 2026-04-01T14:29:18.486 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78966.log 2026-04-01T14:29:18.486 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78931.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.78931.log.gz 2026-04-01T14:29:18.492 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79071.log 2026-04-01T14:29:18.502 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.78966.log: gzip -5 --verbose 85.1% -- -- replaced with /var/log/ceph/ceph-client.admin.78966.log.gz 2026-04-01T14:29:18.502 INFO:teuthology.orchestra.run.vm06.stderr: /var/log/ceph/ceph-client.admin.79170.log 2026-04-01T14:29:18.502 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.79071.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.79071.log.gz 2026-04-01T14:29:18.517 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79966.log 2026-04-01T14:29:18.517 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.79170.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.79170.log.gz 2026-04-01T14:29:18.527 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80000.log 2026-04-01T14:29:18.527 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.79966.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.79966.log.gz 2026-04-01T14:29:18.533 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80034.log 2026-04-01T14:29:18.533 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80000.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.80000.log.gz 2026-04-01T14:29:18.543 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80137.log 2026-04-01T14:29:18.543 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80034.log: 83.1% -- replaced with /var/log/ceph/ceph-client.admin.80034.log.gz 2026-04-01T14:29:18.548 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80235.log 2026-04-01T14:29:18.558 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80137.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.80137.log.gz 2026-04-01T14:29:18.558 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80332.log 2026-04-01T14:29:18.561 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80235.log: 96.3% -- replaced with /var/log/ceph/ceph-client.admin.80235.log.gz 2026-04-01T14:29:18.572 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80366.log 2026-04-01T14:29:18.572 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80332.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.80332.log.gz 2026-04-01T14:29:18.582 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80402.log 2026-04-01T14:29:18.582 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80366.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.80366.log.gz 2026-04-01T14:29:18.588 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80436.log 2026-04-01T14:29:18.588 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80402.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.80402.log.gz 2026-04-01T14:29:18.598 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80470.log 2026-04-01T14:29:18.598 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80436.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.80436.log.gz 2026-04-01T14:29:18.604 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80567.log 2026-04-01T14:29:18.612 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80470.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.80470.log.gz 2026-04-01T14:29:18.613 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80664.log 2026-04-01T14:29:18.621 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80567.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.80567.log.gz 2026-04-01T14:29:18.627 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80761.log 2026-04-01T14:29:18.627 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80664.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.80664.log.gz 2026-04-01T14:29:18.637 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81046.log 2026-04-01T14:29:18.643 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.80761.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81080.log 2026-04-01T14:29:18.643 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81046.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.81046.log.gz 2026-04-01T14:29:18.643 INFO:teuthology.orchestra.run.vm06.stderr: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.80761.log.gz 2026-04-01T14:29:18.657 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81114.log 2026-04-01T14:29:18.657 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81080.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.81080.log.gz 2026-04-01T14:29:18.667 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81217.log 2026-04-01T14:29:18.667 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81114.log: 83.2% -- replaced with /var/log/ceph/ceph-client.admin.81114.log.gz 2026-04-01T14:29:18.672 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81315.log 2026-04-01T14:29:18.682 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81217.log: 84.8% -- replaced with /var/log/ceph/ceph-client.admin.81217.log.gz 2026-04-01T14:29:18.682 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81412.log 2026-04-01T14:29:18.683 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81315.log: 89.1% -- replaced with /var/log/ceph/ceph-client.admin.81315.log.gz 2026-04-01T14:29:18.696 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81446.log 2026-04-01T14:29:18.696 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81412.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.81412.log.gz 2026-04-01T14:29:18.706 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81482.log 2026-04-01T14:29:18.706 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81446.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.81446.log.gz 2026-04-01T14:29:18.711 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81516.log 2026-04-01T14:29:18.720 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81482.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.81482.log.gz 2026-04-01T14:29:18.721 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81550.log 2026-04-01T14:29:18.721 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81516.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.81516.log.gz 2026-04-01T14:29:18.735 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81647.log 2026-04-01T14:29:18.735 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81550.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.81550.log.gz 2026-04-01T14:29:18.750 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81744.log 2026-04-01T14:29:18.750 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81647.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.81647.log.gz 2026-04-01T14:29:18.756 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81841.log 2026-04-01T14:29:18.764 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81744.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.81744.log.gz 2026-04-01T14:29:18.765 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81938.log 2026-04-01T14:29:18.771 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81841.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82035.log 2026-04-01T14:29:18.771 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.81938.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.81841.log.gz 2026-04-01T14:29:18.781 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 85.1% -- replaced with /var/log/ceph/ceph-client.admin.81938.log.gz 2026-04-01T14:29:18.781 INFO:teuthology.orchestra.run.vm06.stderr: --verbose -- /var/log/ceph/ceph-client.admin.82132.log 2026-04-01T14:29:18.781 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82035.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.82035.log.gz 2026-04-01T14:29:18.795 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82229.log 2026-04-01T14:29:18.795 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82132.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.82132.log.gz 2026-04-01T14:29:18.805 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82326.log 2026-04-01T14:29:18.810 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82229.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82423.log 2026-04-01T14:29:18.810 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82326.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.82229.log.gz 2026-04-01T14:29:18.820 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 85.1% -- replaced with /var/log/ceph/ceph-client.admin.82326.log.gz 2026-04-01T14:29:18.820 INFO:teuthology.orchestra.run.vm06.stderr: --verbose -- /var/log/ceph/ceph-client.admin.82520.log 2026-04-01T14:29:18.826 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82423.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82617.log 2026-04-01T14:29:18.826 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82520.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.82423.log.gz 2026-04-01T14:29:18.834 INFO:teuthology.orchestra.run.vm06.stderr: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.82520.log.gz 2026-04-01T14:29:18.835 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82714.log 2026-04-01T14:29:18.843 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82617.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82811.log 2026-04-01T14:29:18.843 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82714.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.82617.log.gz 2026-04-01T14:29:18.850 INFO:teuthology.orchestra.run.vm06.stderr: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82714.log.gz 2026-04-01T14:29:18.851 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82908.log 2026-04-01T14:29:18.856 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82811.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83005.log 2026-04-01T14:29:18.856 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.82908.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82811.log.gz 2026-04-01T14:29:18.865 INFO:teuthology.orchestra.run.vm06.stderr: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82908.log.gz 2026-04-01T14:29:18.866 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83102.log 2026-04-01T14:29:18.866 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83005.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.83005.log.gz 2026-04-01T14:29:18.872 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83199.log 2026-04-01T14:29:18.880 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83102.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.83102.log.gz 2026-04-01T14:29:18.881 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83296.log 2026-04-01T14:29:18.889 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83199.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.83199.log.gz 2026-04-01T14:29:18.895 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83393.log 2026-04-01T14:29:18.895 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83296.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.83296.log.gz 2026-04-01T14:29:18.905 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83490.log 2026-04-01T14:29:18.905 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83393.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.83393.log.gz 2026-04-01T14:29:18.910 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83587.log 2026-04-01T14:29:18.920 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83490.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.83490.log.gz 2026-04-01T14:29:18.920 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83684.log 2026-04-01T14:29:18.920 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83587.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.83587.log.gz 2026-04-01T14:29:18.925 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83781.log 2026-04-01T14:29:18.935 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83684.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.83684.log.gz 2026-04-01T14:29:18.935 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83878.log 2026-04-01T14:29:18.935 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83781.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.83781.log.gz 2026-04-01T14:29:18.941 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.83975.log 2026-04-01T14:29:18.949 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83878.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.83878.log.gz 2026-04-01T14:29:18.950 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84072.log 2026-04-01T14:29:18.958 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.83975.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.83975.log.gz 2026-04-01T14:29:18.965 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84169.log 2026-04-01T14:29:18.965 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84072.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84072.log.gz 2026-04-01T14:29:18.975 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84266.log 2026-04-01T14:29:18.975 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84169.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84169.log.gz 2026-04-01T14:29:18.980 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84363.log 2026-04-01T14:29:18.990 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84266.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.84266.log.gz 2026-04-01T14:29:18.990 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84460.log 2026-04-01T14:29:18.990 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84363.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84363.log.gz 2026-04-01T14:29:18.995 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84557.log 2026-04-01T14:29:19.005 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84460.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84460.log.gz 2026-04-01T14:29:19.005 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84654.log 2026-04-01T14:29:19.005 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84557.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84557.log.gz 2026-04-01T14:29:19.019 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84751.log 2026-04-01T14:29:19.019 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84654.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84654.log.gz 2026-04-01T14:29:19.029 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84848.log 2026-04-01T14:29:19.029 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84751.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84751.log.gz 2026-04-01T14:29:19.036 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.84945.log 2026-04-01T14:29:19.043 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84848.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.84848.log.gz 2026-04-01T14:29:19.044 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85042.log 2026-04-01T14:29:19.053 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.84945.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.84945.log.gz 2026-04-01T14:29:19.058 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85139.log 2026-04-01T14:29:19.058 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85042.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.85042.log.gz 2026-04-01T14:29:19.068 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85236.log 2026-04-01T14:29:19.073 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85139.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85333.log 2026-04-01T14:29:19.073 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85236.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.85139.log.gz 2026-04-01T14:29:19.083 INFO:teuthology.orchestra.run.vm06.stderr: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.85236.log.gz 2026-04-01T14:29:19.098 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85430.log 2026-04-01T14:29:19.098 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85333.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.85333.log.gz 2026-04-01T14:29:19.108 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85527.log 2026-04-01T14:29:19.113 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85430.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85624.log 2026-04-01T14:29:19.113 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85527.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.85430.log.gz 2026-04-01T14:29:19.123 INFO:teuthology.orchestra.run.vm06.stderr: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.85527.log.gz 2026-04-01T14:29:19.128 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85721.log 2026-04-01T14:29:19.133 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85624.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.85624.log.gz 2026-04-01T14:29:19.138 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85818.log 2026-04-01T14:29:19.141 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85721.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.85721.log.gz 2026-04-01T14:29:19.146 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.85915.log 2026-04-01T14:29:19.152 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85818.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.85818.log.gz 2026-04-01T14:29:19.158 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86012.log 2026-04-01T14:29:19.158 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.85915.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.85915.log.gz 2026-04-01T14:29:19.168 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86109.log 2026-04-01T14:29:19.172 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86012.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.86012.log.gz 2026-04-01T14:29:19.175 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86206.log 2026-04-01T14:29:19.183 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86109.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.86109.log.gz 2026-04-01T14:29:19.183 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86303.log 2026-04-01T14:29:19.192 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86206.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.86206.log.gz 2026-04-01T14:29:19.197 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86400.log 2026-04-01T14:29:19.197 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86303.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.86303.log.gz 2026-04-01T14:29:19.205 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86497.log 2026-04-01T14:29:19.214 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86400.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.86400.log.gz 2026-04-01T14:29:19.215 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86594.log 2026-04-01T14:29:19.215 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86497.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.86497.log.gz 2026-04-01T14:29:19.225 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86691.log 2026-04-01T14:29:19.225 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86594.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.86594.log.gz 2026-04-01T14:29:19.230 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86788.log 2026-04-01T14:29:19.240 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86691.log: gzip 84.9% -- replaced with /var/log/ceph/ceph-client.admin.86691.log.gz 2026-04-01T14:29:19.240 INFO:teuthology.orchestra.run.vm06.stderr: -5 --verbose -- /var/log/ceph/ceph-client.admin.86885.log 2026-04-01T14:29:19.240 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86788.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.86788.log.gz 2026-04-01T14:29:19.254 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.86982.log 2026-04-01T14:29:19.254 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86885.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.86885.log.gz 2026-04-01T14:29:19.264 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87079.log 2026-04-01T14:29:19.264 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.86982.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.86982.log.gz 2026-04-01T14:29:19.269 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87176.log 2026-04-01T14:29:19.276 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87079.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87273.log 2026-04-01T14:29:19.278 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87176.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.87079.log.gz 2026-04-01T14:29:19.279 INFO:teuthology.orchestra.run.vm06.stderr: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.87176.log.gz 2026-04-01T14:29:19.294 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87370.log 2026-04-01T14:29:19.294 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87273.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.87273.log.gz 2026-04-01T14:29:19.304 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87467.log 2026-04-01T14:29:19.304 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87370.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.87370.log.gz 2026-04-01T14:29:19.309 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87564.log 2026-04-01T14:29:19.319 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87467.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.87467.log.gz 2026-04-01T14:29:19.319 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.87661.log 2026-04-01T14:29:19.319 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87564.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.87564.log.gz 2026-04-01T14:29:19.333 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.326813.log 2026-04-01T14:29:19.333 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.87661.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.87661.log.gz 2026-04-01T14:29:19.344 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.326847.log 2026-04-01T14:29:19.344 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.326813.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.326813.log.gz 2026-04-01T14:29:19.349 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.326881.log 2026-04-01T14:29:19.358 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.326847.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.326847.log.gz 2026-04-01T14:29:19.359 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.326984.log 2026-04-01T14:29:19.359 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.326881.log: 82.8% -- replaced with /var/log/ceph/ceph-client.admin.326881.log.gz 2026-04-01T14:29:19.373 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.327103.log 2026-04-01T14:29:19.373 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.326984.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.326984.log.gz 2026-04-01T14:29:19.380 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.327200.log 2026-04-01T14:29:19.389 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.327103.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.327234.log 2026-04-01T14:29:19.389 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.327200.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.327200.log.gz 2026-04-01T14:29:19.407 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.327270.log 2026-04-01T14:29:19.407 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.327234.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.327234.log.gz 2026-04-01T14:29:19.423 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.327304.log 2026-04-01T14:29:19.423 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.327270.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.327270.log.gz 2026-04-01T14:29:19.440 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.542682.log 2026-04-01T14:29:19.440 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.327304.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.327304.log.gz 2026-04-01T14:29:19.456 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.542879.log 2026-04-01T14:29:19.456 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.542682.log: 8.2% -- replaced with /var/log/ceph/ceph-client.0.542682.log.gz 2026-04-01T14:29:19.471 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.542955.log 2026-04-01T14:29:19.471 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.0.542879.log: 8.2% -- replaced with /var/log/ceph/ceph-client.0.542879.log.gz 2026-04-01T14:29:19.487 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ceph-client.admin.542955.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.542955.log.gz 2026-04-01T14:29:19.809 INFO:teuthology.orchestra.run.vm06.stderr: 86.8% -- replaced with /var/log/ceph/ceph-client.admin.327103.log.gz 2026-04-01T14:29:19.836 INFO:teuthology.orchestra.run.vm06.stderr: 92.3% -- replaced with /var/log/ceph/ops-log-ceph-client.0.log.gz 2026-04-01T14:29:49.761 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-01T14:29:49.761 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-osd.1.log.gz: No space left on device 2026-04-01T14:29:49.761 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-01T14:29:49.761 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-osd.3.log.gz: No space left on device 2026-04-01T14:29:49.764 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-01T14:29:49.764 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-osd.2.log.gz: No space left on device 2026-04-01T14:29:49.765 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-01T14:29:49.765 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/rgw.ceph.client.0.log.gz: No space left on device 2026-04-01T14:30:16.603 INFO:teuthology.orchestra.run.vm06.stderr: 93.3% -- replaced with /var/log/ceph/ceph-osd.0.log.gz 2026-04-01T14:30:16.604 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-01T14:30:16.604 INFO:teuthology.orchestra.run.vm06.stderr:real 0m59.462s 2026-04-01T14:30:16.604 INFO:teuthology.orchestra.run.vm06.stderr:user 2m22.952s 2026-04-01T14:30:16.604 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m9.030s 2026-04-01T14:30:16.604 DEBUG:teuthology.orchestra.run:got remote process result: 123 2026-04-01T14:30:16.604 ERROR:teuthology.run_tasks:Manager failed: ceph Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 1181, in cluster yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 1996, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/ceph.py", line 263, in ceph_log run.wait( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 123: "time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose --" 2026-04-01T14:30:16.605 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-04-01T14:30:16.608 ERROR:teuthology.contextutil:Saw exception from nested tasks Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/__init__.py", line 644, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' 2026-04-01T14:30:16.609 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-04-01T14:30:16.609 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-04-01T14:30:16.647 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-04-01T14:30:16.649 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-04-01T14:30:16.707 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-04-01T14:30:16.707 DEBUG:teuthology.orchestra.run.vm06:> 2026-04-01T14:30:16.707 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-04-01T14:30:16.707 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-04-01T14:30:16.707 DEBUG:teuthology.orchestra.run.vm06:> done 2026-04-01T14:30:16.715 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-04-01T14:30:16.715 DEBUG:teuthology.orchestra.run.vm08:> 2026-04-01T14:30:16.715 DEBUG:teuthology.orchestra.run.vm08:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-04-01T14:30:16.715 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y remove $d || true 2026-04-01T14:30:16.715 DEBUG:teuthology.orchestra.run.vm08:> done 2026-04-01T14:30:16.723 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-04-01T14:30:16.723 DEBUG:teuthology.orchestra.run.vm09:> 2026-04-01T14:30:16.723 DEBUG:teuthology.orchestra.run.vm09:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-04-01T14:30:16.723 DEBUG:teuthology.orchestra.run.vm09:> sudo yum -y remove $d || true 2026-04-01T14:30:16.723 DEBUG:teuthology.orchestra.run.vm09:> done 2026-04-01T14:30:16.852 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:16.919 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repo Size 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 103 M 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout: mailcap noarch 2.1.49-5.el9.0.2 @baseos 78 k 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:Remove 2 Packages 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 103 M 2026-04-01T14:30:16.920 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-04-01T14:30:16.922 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-04-01T14:30:16.923 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-04-01T14:30:16.934 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-04-01T14:30:16.935 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-04-01T14:30:16.966 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-04-01T14:30:16.991 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:16.991 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:16.991 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-01T14:30:16.991 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-04-01T14:30:16.991 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-04-01T14:30:16.991 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:16.998 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.000 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.012 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.027 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-01T14:30:17.037 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repo Size 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 103 M 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9.0.2 @baseos 78 k 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:Remove 2 Packages 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 103 M 2026-04-01T14:30:17.038 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-04-01T14:30:17.042 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-04-01T14:30:17.042 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-04-01T14:30:17.056 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-04-01T14:30:17.056 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-04-01T14:30:17.091 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-04-01T14:30:17.101 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-01T14:30:17.101 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.115 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.115 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:17.115 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-01T14:30:17.115 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-04-01T14:30:17.115 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-04-01T14:30:17.115 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.119 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.133 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.143 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.148 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-01T14:30:17.187 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-01T14:30:17.188 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.188 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-04-01T14:30:17.188 INFO:teuthology.orchestra.run.vm09.stdout: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:17.188 INFO:teuthology.orchestra.run.vm09.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-01T14:30:17.188 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.188 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-04-01T14:30:17.282 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.419 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 365 M 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:17.461 INFO:teuthology.orchestra.run.vm09.stdout:Remove 3 Packages 2026-04-01T14:30:17.462 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.462 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 366 M 2026-04-01T14:30:17.462 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-04-01T14:30:17.464 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-04-01T14:30:17.464 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-04-01T14:30:17.484 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-04-01T14:30:17.484 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-04-01T14:30:17.561 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.566 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-04-01T14:30:17.574 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/3 2026-04-01T14:30:17.578 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-01T14:30:17.578 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 1/2 2026-04-01T14:30:17.578 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-01T14:30:17.594 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-01T14:30:17.661 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-01T14:30:17.661 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/3 2026-04-01T14:30:17.661 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.679 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-04-01T14:30:17.702 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout: ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 socat-1.7.4.1-8.el9.x86_64 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.715 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-04-01T14:30:17.849 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 365 M 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:Remove 3 Packages 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:17.906 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 366 M 2026-04-01T14:30:17.907 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-04-01T14:30:17.909 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-04-01T14:30:17.909 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-04-01T14:30:17.931 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-04-01T14:30:17.931 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-04-01T14:30:17.936 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 0 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 6.8 M 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 19 M 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-04-01T14:30:17.937 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-04-01T14:30:17.938 INFO:teuthology.orchestra.run.vm09.stdout:Remove 8 Packages 2026-04-01T14:30:17.938 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:17.938 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 28 M 2026-04-01T14:30:17.938 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-04-01T14:30:17.940 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-04-01T14:30:17.940 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-04-01T14:30:17.963 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-04-01T14:30:17.963 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-04-01T14:30:17.985 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:17.988 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-04-01T14:30:17.995 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/3 2026-04-01T14:30:17.998 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-01T14:30:18.001 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-04-01T14:30:18.006 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/8 2026-04-01T14:30:18.010 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-04-01T14:30:18.013 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-04-01T14:30:18.014 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-01T14:30:18.015 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-04-01T14:30:18.019 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-04-01T14:30:18.021 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-04-01T14:30:18.083 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 7/8 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/3 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-01T14:30:18.084 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 7/8 2026-04-01T14:30:18.092 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 7/8 2026-04-01T14:30:18.111 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 8/8 2026-04-01T14:30:18.111 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:18.111 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-01T14:30:18.111 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-04-01T14:30:18.111 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-04-01T14:30:18.111 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.113 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 8/8 2026-04-01T14:30:18.130 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 socat-1.7.4.1-8.el9.x86_64 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.134 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 8/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 3/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-04-01T14:30:18.198 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout:Removed: 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: lua-5.4.4-4.el9.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: unzip-6.0-59.el9.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: zip-3.0-35.el9.x86_64 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.243 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-04-01T14:30:18.269 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.328 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 0 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 6.8 M 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 19 M 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:Remove 8 Packages 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 28 M 2026-04-01T14:30:18.329 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-04-01T14:30:18.332 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-04-01T14:30:18.332 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-04-01T14:30:18.356 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-04-01T14:30:18.356 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-04-01T14:30:18.394 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-04-01T14:30:18.400 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/8 2026-04-01T14:30:18.403 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-04-01T14:30:18.405 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-04-01T14:30:18.408 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-04-01T14:30:18.411 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-04-01T14:30:18.412 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-04-01T14:30:18.417 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 7/8 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.440 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 7/8 2026-04-01T14:30:18.447 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 7/8 2026-04-01T14:30:18.465 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-04-01T14:30:18.471 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 8/8 2026-04-01T14:30:18.471 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:18.471 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-01T14:30:18.471 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-04-01T14:30:18.471 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-04-01T14:30:18.471 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout:=================================================================================================================== 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout:=================================================================================================================== 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout:Removing: 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-base x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 24 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout:Removing dependent packages: 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 447 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 2.9 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 938 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-dashboard noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 148 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 66 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-rook noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 567 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-osd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 54 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-volume noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 1.4 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: rbd-mirror x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 11 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout:Removing unused dependencies: 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: c-ares x86_64 1.19.1-2.el9_4 @baseos 279 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 98 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-grafana-dashboards noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 990 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 60 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-mgr-modules-core noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 1.6 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-prometheus-alerts noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 57 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ceph-selinux x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 138 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: cryptsetup x86_64 2.7.2-4.el9 @baseos 722 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 @appstream 68 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 @appstream 11 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 @appstream 39 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: libcephsqlite x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 409 k 2026-04-01T14:30:18.472 INFO:teuthology.orchestra.run.vm09.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: libgfortran x86_64 11.5.0-11.el9 @baseos 2.8 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: libquadmath x86_64 11.5.0-11.el9 @baseos 330 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: libradosstriper1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 792 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: libxslt x86_64 1.1.34-13.el9_6 @appstream 751 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli x86_64 2.13-1.el9 @baseos 6.8 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: protobuf x86_64 3.14.0-17.el9_7 @appstream 3.5 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 @crb 2.9 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 816 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-chardet noarch 4.0.0-5.el9 @77d52b2cce1347aa9f3fc60d8b93d222 1.4 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-cheroot noarch 10.0.1-5.el9 @epel 682 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-cherrypy noarch 18.10.0-5.el9 @epel 1.0 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 @baseos 4.5 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-devel x86_64 3.9.23-2.el9 @appstream 765 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-idna noarch 2.10-7.el9_4.1 @77d52b2cce1347aa9f3fc60d8b93d222 513 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 @appstream 1.1 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpatch noarch 1.21-16.el9 @0d57cd3fe20446e8b1c08da162742194 55 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-jsonpointer noarch 2.0-4.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 34 k 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-04-01T14:30:18.473 INFO:teuthology.orchestra.run.vm09.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 @appstream 30 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 @appstream 1.7 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-oauthlib noarch 3.1.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 888 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-ply noarch 3.11-14.el9.0.1 @baseos 430 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-prettytable noarch 0.7.2-27.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 166 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-protobuf noarch 3.14.0-17.el9_7 @appstream 1.4 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 @appstream 622 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 @appstream 1.0 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 @baseos 635 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pysocks noarch 1.7.1-12.el9.0.1 @77d52b2cce1347aa9f3fc60d8b93d222 88 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-pytz noarch 2021.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 176 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests noarch 2.25.1-10.el9_6 @baseos 405 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 72 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-toml noarch 0.10.2-6.el9.0.1 @appstream 99 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 @baseos 746 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: qatlib x86_64 24.09.0-1.el9 @appstream 588 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: qatlib-service x86_64 24.09.0-1.el9 @appstream 64 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: smartmontools x86_64 1:7.2-9.el9 @baseos 1.9 M 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout:=================================================================================================================== 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout:Remove 111 Packages 2026-04-01T14:30:18.474 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.475 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 8/8 2026-04-01T14:30:18.475 INFO:teuthology.orchestra.run.vm09.stdout:Freed space: 687 M 2026-04-01T14:30:18.475 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-04-01T14:30:18.499 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-04-01T14:30:18.499 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-04-01T14:30:18.569 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 8/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 1/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 3/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-04-01T14:30:18.571 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-04-01T14:30:18.620 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-04-01T14:30:18.620 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout:Removed: 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.621 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-04-01T14:30:18.707 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.779 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-04-01T14:30:18.779 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 1/111 2026-04-01T14:30:18.788 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 1/111 2026-04-01T14:30:18.810 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 2/111 2026-04-01T14:30:18.810 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:18.810 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-01T14:30:18.810 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-04-01T14:30:18.810 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-04-01T14:30:18.810 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:18.811 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 2/111 2026-04-01T14:30:18.824 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-04-01T14:30:18.826 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 2/111 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout:=================================================================================================================== 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout:=================================================================================================================== 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout:Removing: 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 24 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout:Removing dependent packages: 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 447 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 2.9 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 938 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 148 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 66 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 567 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 54 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 1.4 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 11 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout:Removing unused dependencies: 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: c-ares x86_64 1.19.1-2.el9_4 @baseos 279 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 98 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 990 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 60 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 1.6 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph-noarch 57 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 138 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.7.2-4.el9 @baseos 722 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 @appstream 68 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 @appstream 11 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 @appstream 39 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 409 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-11.el9 @baseos 2.8 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-11.el9 @baseos 330 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 792 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-13.el9_6 @appstream 751 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.13-1.el9 @baseos 6.8 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9_7 @appstream 3.5 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 @crb 2.9 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-04-01T14:30:18.831 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:20.2.0-9.ge2ce8426bc5.el9.clyso @ceph 816 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-chardet noarch 4.0.0-5.el9 @77d52b2cce1347aa9f3fc60d8b93d222 1.4 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-5.el9 @epel 682 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.10.0-5.el9 @epel 1.0 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 @baseos 4.5 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.23-2.el9 @appstream 765 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-idna noarch 2.10-7.el9_4.1 @77d52b2cce1347aa9f3fc60d8b93d222 513 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 @appstream 1.1 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpatch noarch 1.21-16.el9 @0d57cd3fe20446e8b1c08da162742194 55 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-jsonpointer noarch 2.0-4.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 34 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 @appstream 30 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 @appstream 1.7 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-oauthlib noarch 3.1.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 888 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9.0.1 @baseos 430 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-prettytable noarch 0.7.2-27.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 166 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9_7 @appstream 1.4 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 @appstream 622 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 @appstream 1.0 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 @baseos 635 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pysocks noarch 1.7.1-12.el9.0.1 @77d52b2cce1347aa9f3fc60d8b93d222 88 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-pytz noarch 2021.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 176 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9_6 @baseos 405 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 72 M 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9.0.1 @appstream 99 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 @baseos 746 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 24.09.0-1.el9 @appstream 588 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 24.09.0-1.el9 @appstream 64 k 2026-04-01T14:30:18.832 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools x86_64 1:7.2-9.el9 @baseos 1.9 M 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout:=================================================================================================================== 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout:Remove 111 Packages 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout:Freed space: 687 M 2026-04-01T14:30:18.833 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-04-01T14:30:18.842 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.858 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-04-01T14:30:18.858 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-04-01T14:30:18.928 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 3/111 2026-04-01T14:30:18.928 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 4/111 2026-04-01T14:30:18.975 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 4/111 2026-04-01T14:30:18.976 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:18.982 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/111 2026-04-01T14:30:18.982 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/111 2026-04-01T14:30:18.989 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-04-01T14:30:18.990 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-04-01T14:30:18.998 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/111 2026-04-01T14:30:19.007 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cherrypy-18.10.0-5.el9.noarch 7/111 2026-04-01T14:30:19.012 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cheroot-10.0.1-5.el9.noarch 8/111 2026-04-01T14:30:19.021 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/111 2026-04-01T14:30:19.027 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/111 2026-04-01T14:30:19.055 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 11/111 2026-04-01T14:30:19.055 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.055 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-01T14:30:19.055 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-04-01T14:30:19.055 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-04-01T14:30:19.055 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.061 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 11/111 2026-04-01T14:30:19.071 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 11/111 2026-04-01T14:30:19.089 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 12/111 2026-04-01T14:30:19.089 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.089 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-01T14:30:19.089 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.099 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 12/111 2026-04-01T14:30:19.110 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 12/111 2026-04-01T14:30:19.113 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/111 2026-04-01T14:30:19.116 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.119 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/111 2026-04-01T14:30:19.124 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jinja2-2.11.3-8.el9_5.noarch 15/111 2026-04-01T14:30:19.156 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/111 2026-04-01T14:30:19.164 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-04-01T14:30:19.164 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 1/111 2026-04-01T14:30:19.164 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 17/111 2026-04-01T14:30:19.167 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 18/111 2026-04-01T14:30:19.172 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-9.ge2ce8426bc5.el9.clyso. 1/111 2026-04-01T14:30:19.177 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 19/111 2026-04-01T14:30:19.188 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 20/111 2026-04-01T14:30:19.188 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 21/111 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 2/111 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:19.192 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 2/111 2026-04-01T14:30:19.196 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 21/111 2026-04-01T14:30:19.206 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 2/111 2026-04-01T14:30:19.244 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.292 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 22/111 2026-04-01T14:30:19.302 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-modules-core-2:20.2.0-9.ge2ce8426bc5.el 3/111 2026-04-01T14:30:19.302 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 4/111 2026-04-01T14:30:19.309 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 23/111 2026-04-01T14:30:19.316 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 24/111 2026-04-01T14:30:19.320 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 25/111 2026-04-01T14:30:19.324 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-9.ge2ce8426bc5.el9.c 4/111 2026-04-01T14:30:19.331 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/111 2026-04-01T14:30:19.331 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/111 2026-04-01T14:30:19.333 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-01T14:30:19.334 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-04-01T14:30:19.334 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.335 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-01T14:30:19.344 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-9.ge2ce8426bc5.el9.cly 6/111 2026-04-01T14:30:19.352 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cherrypy-18.10.0-5.el9.noarch 7/111 2026-04-01T14:30:19.357 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cheroot-10.0.1-5.el9.noarch 8/111 2026-04-01T14:30:19.364 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-01T14:30:19.366 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/111 2026-04-01T14:30:19.369 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 27/111 2026-04-01T14:30:19.371 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/111 2026-04-01T14:30:19.372 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 28/111 2026-04-01T14:30:19.376 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.386 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 29/111 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cryptography-36.0.1-5.el9_6.x86_64 30/111 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 11/111 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-04-01T14:30:19.392 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:19.396 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 11/111 2026-04-01T14:30:19.396 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : protobuf-compiler-3.14.0-17.el9_7.x86_64 31/111 2026-04-01T14:30:19.399 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 32/111 2026-04-01T14:30:19.406 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_6 11/111 2026-04-01T14:30:19.422 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 12/111 2026-04-01T14:30:19.422 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.422 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-01T14:30:19.422 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:19.424 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 33/111 2026-04-01T14:30:19.425 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.425 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-01T14:30:19.425 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-04-01T14:30:19.425 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-04-01T14:30:19.425 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.425 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 33/111 2026-04-01T14:30:19.431 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 12/111 2026-04-01T14:30:19.435 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 33/111 2026-04-01T14:30:19.439 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 34/111 2026-04-01T14:30:19.440 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:20.2.0-9.ge2ce8426bc5.el9.clyso.no 12/111 2026-04-01T14:30:19.442 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 35/111 2026-04-01T14:30:19.443 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/111 2026-04-01T14:30:19.445 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 36/111 2026-04-01T14:30:19.448 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 37/111 2026-04-01T14:30:19.448 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/111 2026-04-01T14:30:19.451 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 38/111 2026-04-01T14:30:19.454 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jinja2-2.11.3-8.el9_5.noarch 15/111 2026-04-01T14:30:19.454 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 39/111 2026-04-01T14:30:19.454 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 40/111 2026-04-01T14:30:19.484 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/111 2026-04-01T14:30:19.492 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 17/111 2026-04-01T14:30:19.495 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 18/111 2026-04-01T14:30:19.504 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 19/111 2026-04-01T14:30:19.511 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 40/111 2026-04-01T14:30:19.512 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 20/111 2026-04-01T14:30:19.513 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 21/111 2026-04-01T14:30:19.514 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.521 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-9.ge2ce84 21/111 2026-04-01T14:30:19.522 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 41/111 2026-04-01T14:30:19.527 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 42/111 2026-04-01T14:30:19.536 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-requests-2.25.1-10.el9_6.noarch 43/111 2026-04-01T14:30:19.541 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 44/111 2026-04-01T14:30:19.551 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 45/111 2026-04-01T14:30:19.558 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 46/111 2026-04-01T14:30:19.562 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-urllib3-1.26.5-6.el9_7.1.noarch 47/111 2026-04-01T14:30:19.568 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 48/111 2026-04-01T14:30:19.619 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 49/111 2026-04-01T14:30:19.623 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 22/111 2026-04-01T14:30:19.632 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-numpy-1:1.23.5-2.el9_7.x86_64 50/111 2026-04-01T14:30:19.636 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 51/111 2026-04-01T14:30:19.642 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 52/111 2026-04-01T14:30:19.643 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 23/111 2026-04-01T14:30:19.644 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 53/111 2026-04-01T14:30:19.648 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : libgfortran-11.5.0-11.el9.x86_64 54/111 2026-04-01T14:30:19.651 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 24/111 2026-04-01T14:30:19.651 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 55/111 2026-04-01T14:30:19.655 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.656 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 25/111 2026-04-01T14:30:19.671 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-01T14:30:19.671 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-04-01T14:30:19.671 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:19.672 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-01T14:30:19.673 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 56/111 2026-04-01T14:30:19.673 INFO:teuthology.orchestra.run.vm09.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.673 INFO:teuthology.orchestra.run.vm09.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-01T14:30:19.673 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.674 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 56/111 2026-04-01T14:30:19.682 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 56/111 2026-04-01T14:30:19.684 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 57/111 2026-04-01T14:30:19.687 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : flexiblas-3.0.4-8.el9.0.1.x86_64 58/111 2026-04-01T14:30:19.690 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ply-3.11-14.el9.0.1.noarch 59/111 2026-04-01T14:30:19.693 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-idna-2.10-7.el9_4.1.noarch 60/111 2026-04-01T14:30:19.698 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pysocks-1.7.1-12.el9.0.1.noarch 61/111 2026-04-01T14:30:19.703 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyasn1-0.4.8-7.el9_7.noarch 62/111 2026-04-01T14:30:19.708 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-01T14:30:19.709 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 63/111 2026-04-01T14:30:19.714 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 27/111 2026-04-01T14:30:19.716 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 28/111 2026-04-01T14:30:19.718 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 64/111 2026-04-01T14:30:19.725 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 65/111 2026-04-01T14:30:19.728 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 66/111 2026-04-01T14:30:19.730 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 29/111 2026-04-01T14:30:19.731 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 67/111 2026-04-01T14:30:19.733 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 68/111 2026-04-01T14:30:19.736 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 69/111 2026-04-01T14:30:19.736 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cryptography-36.0.1-5.el9_6.x86_64 30/111 2026-04-01T14:30:19.739 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 70/111 2026-04-01T14:30:19.740 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : protobuf-compiler-3.14.0-17.el9_7.x86_64 31/111 2026-04-01T14:30:19.743 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 32/111 2026-04-01T14:30:19.743 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-toml-0.10.2-6.el9.0.1.noarch 71/111 2026-04-01T14:30:19.746 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 72/111 2026-04-01T14:30:19.749 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 73/111 2026-04-01T14:30:19.758 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/111 2026-04-01T14:30:19.762 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 33/111 2026-04-01T14:30:19.762 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:19.762 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-01T14:30:19.762 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-04-01T14:30:19.762 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-04-01T14:30:19.762 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:19.763 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 33/111 2026-04-01T14:30:19.763 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-devel-3.9.23-2.el9.x86_64 75/111 2026-04-01T14:30:19.765 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-jsonpointer-2.0-4.el9.0.1.noarch 76/111 2026-04-01T14:30:19.769 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/111 2026-04-01T14:30:19.771 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86 33/111 2026-04-01T14:30:19.772 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 78/111 2026-04-01T14:30:19.774 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 79/111 2026-04-01T14:30:19.776 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 34/111 2026-04-01T14:30:19.779 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 35/111 2026-04-01T14:30:19.780 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 80/111 2026-04-01T14:30:19.782 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 36/111 2026-04-01T14:30:19.784 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-protobuf-3.14.0-17.el9_7.noarch 81/111 2026-04-01T14:30:19.785 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 37/111 2026-04-01T14:30:19.787 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.788 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 82/111 2026-04-01T14:30:19.789 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 38/111 2026-04-01T14:30:19.791 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 83/111 2026-04-01T14:30:19.792 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 39/111 2026-04-01T14:30:19.792 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 40/111 2026-04-01T14:30:19.793 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 84/111 2026-04-01T14:30:19.795 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 85/111 2026-04-01T14:30:19.814 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 86/111 2026-04-01T14:30:19.814 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-04-01T14:30:19.814 INFO:teuthology.orchestra.run.vm09.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-04-01T14:30:19.814 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.822 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 86/111 2026-04-01T14:30:19.823 INFO:teuthology.orchestra.run.vm09.stdout:warning: file /etc/logrotate.d/ceph: remove failed: No such file or directory 2026-04-01T14:30:19.823 INFO:teuthology.orchestra.run.vm09.stdout: 2026-04-01T14:30:19.853 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-9.ge2ce8426bc5.el9.c 40/111 2026-04-01T14:30:19.860 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 86/111 2026-04-01T14:30:19.860 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 87/111 2026-04-01T14:30:19.862 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 41/111 2026-04-01T14:30:19.866 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 42/111 2026-04-01T14:30:19.874 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 87/111 2026-04-01T14:30:19.877 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-requests-2.25.1-10.el9_6.noarch 43/111 2026-04-01T14:30:19.880 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 88/111 2026-04-01T14:30:19.883 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 89/111 2026-04-01T14:30:19.883 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 44/111 2026-04-01T14:30:19.886 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : python3-prettytable-0.7.2-27.el9.0.1.noarch 90/111 2026-04-01T14:30:19.886 INFO:teuthology.orchestra.run.vm09.stdout: Erasing : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 91/111 2026-04-01T14:30:19.895 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 45/111 2026-04-01T14:30:19.903 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 46/111 2026-04-01T14:30:19.908 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-urllib3-1.26.5-6.el9_7.1.noarch 47/111 2026-04-01T14:30:19.914 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 48/111 2026-04-01T14:30:19.920 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:19.965 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 49/111 2026-04-01T14:30:19.979 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-numpy-1:1.23.5-2.el9_7.x86_64 50/111 2026-04-01T14:30:19.983 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 51/111 2026-04-01T14:30:19.986 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 52/111 2026-04-01T14:30:19.989 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 53/111 2026-04-01T14:30:19.993 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : libgfortran-11.5.0-11.el9.x86_64 54/111 2026-04-01T14:30:19.996 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 55/111 2026-04-01T14:30:20.017 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 56/111 2026-04-01T14:30:20.017 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-01T14:30:20.017 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-01T14:30:20.017 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:20.018 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 56/111 2026-04-01T14:30:20.027 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-9.ge2ce8426 56/111 2026-04-01T14:30:20.029 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 57/111 2026-04-01T14:30:20.031 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : flexiblas-3.0.4-8.el9.0.1.x86_64 58/111 2026-04-01T14:30:20.034 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ply-3.11-14.el9.0.1.noarch 59/111 2026-04-01T14:30:20.037 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-idna-2.10-7.el9_4.1.noarch 60/111 2026-04-01T14:30:20.043 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pysocks-1.7.1-12.el9.0.1.noarch 61/111 2026-04-01T14:30:20.048 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyasn1-0.4.8-7.el9_7.noarch 62/111 2026-04-01T14:30:20.052 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:20.054 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 63/111 2026-04-01T14:30:20.064 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 64/111 2026-04-01T14:30:20.071 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 65/111 2026-04-01T14:30:20.074 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 66/111 2026-04-01T14:30:20.077 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 67/111 2026-04-01T14:30:20.079 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 68/111 2026-04-01T14:30:20.082 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 69/111 2026-04-01T14:30:20.085 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 70/111 2026-04-01T14:30:20.087 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-toml-0.10.2-6.el9.0.1.noarch 71/111 2026-04-01T14:30:20.090 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 72/111 2026-04-01T14:30:20.093 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 73/111 2026-04-01T14:30:20.101 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/111 2026-04-01T14:30:20.105 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-devel-3.9.23-2.el9.x86_64 75/111 2026-04-01T14:30:20.108 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-jsonpointer-2.0-4.el9.0.1.noarch 76/111 2026-04-01T14:30:20.112 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/111 2026-04-01T14:30:20.115 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 78/111 2026-04-01T14:30:20.118 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 79/111 2026-04-01T14:30:20.123 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 80/111 2026-04-01T14:30:20.127 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-protobuf-3.14.0-17.el9_7.noarch 81/111 2026-04-01T14:30:20.131 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 82/111 2026-04-01T14:30:20.134 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 83/111 2026-04-01T14:30:20.135 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-grafana-dashboards-2:20.2.0-9.ge2ce8426bc5. 84/111 2026-04-01T14:30:20.137 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-prometheus-alerts-2:20.2.0-9.ge2ce8426bc5.e 85/111 2026-04-01T14:30:20.157 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 86/111 2026-04-01T14:30:20.157 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-04-01T14:30:20.157 INFO:teuthology.orchestra.run.vm06.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-04-01T14:30:20.157 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:20.164 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 86/111 2026-04-01T14:30:20.164 INFO:teuthology.orchestra.run.vm06.stdout:warning: file /etc/logrotate.d/ceph: remove failed: No such file or directory 2026-04-01T14:30:20.164 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-01T14:30:20.183 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:20.190 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x86_ 86/111 2026-04-01T14:30:20.190 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 87/111 2026-04-01T14:30:20.201 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-04-01T14:30:20.314 INFO:teuthology.orchestra.run.vm08.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-01T14:30:20.332 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:30:20.333 ERROR:teuthology.run_tasks:Manager failed: install Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/__init__.py", line 220, in install yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/__init__.py", line 644, in task yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/__init__.py", line 640, in task with contextutil.nested( File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/__init__.py", line 222, in install remove_packages(ctx, config, package_list) File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/__init__.py", line 103, in remove_packages with parallel() as p: File "/home/teuthos/kshtsk/teuthology/teuthology/parallel.py", line 84, in __exit__ for result in self: File "/home/teuthos/kshtsk/teuthology/teuthology/parallel.py", line 98, in __next__ resurrect_traceback(result) File "/home/teuthos/kshtsk/teuthology/teuthology/parallel.py", line 30, in resurrect_traceback raise exc.exc_info[1] File "/home/teuthos/kshtsk/teuthology/teuthology/parallel.py", line 23, in capture_traceback return func(*args, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/task/install/rpm.py", line 43, in _remove remote.run(args='sudo yum clean all') File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm08 with status 1: 'sudo yum clean all' 2026-04-01T14:30:20.333 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-04-01T14:30:20.336 INFO:teuthology.task.clock:Checking final clock skew... 2026-04-01T14:30:20.336 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-01T14:30:20.337 DEBUG:teuthology.orchestra.run.vm08:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-01T14:30:20.339 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-01T14:30:20.352 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-04-01T14:30:20.353 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-04-01T14:30:20.354 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm08.stdout:^* node-3.infogral.is 2 8 377 26 +744us[ +723us] +/- 14ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm08.stdout:^+ ntp.b-ite.de 2 6 377 26 +796us[ +796us] +/- 18ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm08.stdout:^+ 185.13.148.71 2 7 377 27 +673us[ +651us] +/- 18ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm08.stdout:^+ byggvir.de 2 7 377 94 -2940us[-2962us] +/- 22ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm06.stdout:^+ ntp.b-ite.de 2 6 377 27 +835us[ +835us] +/- 18ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm06.stdout:^+ 185.13.148.71 2 7 377 27 +606us[ +606us] +/- 18ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm06.stdout:^+ byggvir.de 2 7 377 92 -2937us[-2939us] +/- 22ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm06.stdout:^* node-3.infogral.is 2 8 377 28 +617us[ +617us] +/- 14ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm09.stdout:^+ ntp.b-ite.de 2 8 377 94 +984us[ +985us] +/- 18ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm09.stdout:^+ 185.13.148.71 2 7 377 31 +648us[ +648us] +/- 18ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm09.stdout:^+ byggvir.de 2 6 377 27 -2906us[-2906us] +/- 23ms 2026-04-01T14:30:20.485 INFO:teuthology.orchestra.run.vm09.stdout:^* node-3.infogral.is 2 7 377 93 +620us[ +621us] +/- 14ms 2026-04-01T14:30:20.486 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-04-01T14:30:20.488 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-04-01T14:30:20.491 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-04-01T14:30:20.494 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-04-01T14:30:20.496 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-04-01T14:30:20.499 INFO:teuthology.task.internal:Duration was 2471.015863 seconds 2026-04-01T14:30:20.499 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-04-01T14:30:20.501 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-04-01T14:30:20.501 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-04-01T14:30:20.527 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-04-01T14:30:20.529 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-04-01T14:30:20.568 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-01T14:30:20.569 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-01T14:30:20.570 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-01T14:30:20.938 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-04-01T14:30:20.938 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-04-01T14:30:20.938 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-04-01T14:30:20.962 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm08.local 2026-04-01T14:30:20.963 DEBUG:teuthology.orchestra.run.vm08:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-04-01T14:30:20.985 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-04-01T14:30:20.985 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-04-01T14:30:21.009 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-04-01T14:30:21.009 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-04-01T14:30:21.011 DEBUG:teuthology.orchestra.run.vm08:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-04-01T14:30:21.027 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-04-01T14:30:21.043 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: /home/ubuntu/cephtest/archive/syslog/journalctl.log: No space left on device 2026-04-01T14:30:21.050 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:30:21.050 ERROR:teuthology.run_tasks:Manager failed: internal.syslog Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/task/internal/syslog.py", line 76, in syslog yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/kshtsk/teuthology/teuthology/task/internal/syslog.py", line 163, in syslog run.wait( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm08 with status 1: 'sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log' 2026-04-01T14:30:21.051 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-04-01T14:30:21.053 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-04-01T14:30:21.053 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-04-01T14:30:21.116 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-04-01T14:30:21.142 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-04-01T14:30:21.170 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-04-01T14:30:21.173 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-04-01T14:30:21.174 DEBUG:teuthology.orchestra.run.vm08:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-04-01T14:30:21.184 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-04-01T14:30:21.187 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x8 87/111 2026-04-01T14:30:21.194 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 88/111 2026-04-01T14:30:21.197 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-ceph-common-2:20.2.0-9.ge2ce8426bc5.el9. 89/111 2026-04-01T14:30:21.200 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : python3-prettytable-0.7.2-27.el9.0.1.noarch 90/111 2026-04-01T14:30:21.200 INFO:teuthology.orchestra.run.vm06.stdout: Erasing : ceph-selinux-2:20.2.0-9.ge2ce8426bc5.el9.clyso.x 91/111 2026-04-01T14:30:21.202 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-04-01T14:30:21.209 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = core 2026-04-01T14:30:21.240 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-04-01T14:30:21.254 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-04-01T14:30:21.275 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:30:21.275 DEBUG:teuthology.orchestra.run.vm08:> test -e /home/ubuntu/cephtest/archive/coredump 2026-04-01T14:30:21.292 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:30:21.292 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-04-01T14:30:21.312 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:30:21.312 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-04-01T14:30:21.315 INFO:teuthology.task.internal:Transferring archived files... 2026-04-01T14:30:21.320 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/remote/vm06 2026-04-01T14:30:21.320 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-04-01T14:30:21.346 DEBUG:teuthology.misc:Transferring archived files from vm08:/home/ubuntu/cephtest/archive to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/remote/vm08 2026-04-01T14:30:21.347 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-04-01T14:30:21.373 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/supriti-2026-04-01_13:45:16-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/4802/remote/vm09 2026-04-01T14:30:21.373 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-04-01T14:30:21.543 INFO:teuthology.task.internal:Removing archive directory... 2026-04-01T14:30:21.543 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-04-01T14:30:21.544 DEBUG:teuthology.orchestra.run.vm08:> rm -rf -- /home/ubuntu/cephtest/archive 2026-04-01T14:30:21.546 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-04-01T14:30:21.602 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-04-01T14:30:21.605 INFO:teuthology.task.internal:Not uploading archives. 2026-04-01T14:30:21.605 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-04-01T14:30:21.608 INFO:teuthology.task.internal:Tidying up after the test... 2026-04-01T14:30:21.608 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-04-01T14:30:21.610 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-04-01T14:30:21.612 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-04-01T14:30:21.627 INFO:teuthology.orchestra.run.vm06.stdout: 83886506 0 drwxr-xr-x 3 ubuntu ubuntu 23 Apr 1 14:30 /home/ubuntu/cephtest 2026-04-01T14:30:21.627 INFO:teuthology.orchestra.run.vm06.stdout: 58721128 0 drwxr-xr-x 2 ubuntu ubuntu 6 Apr 1 13:54 /home/ubuntu/cephtest/ceph.data 2026-04-01T14:30:21.628 INFO:teuthology.orchestra.run.vm06.stderr:rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty 2026-04-01T14:30:21.629 INFO:teuthology.orchestra.run.vm08.stdout: 83886506 0 drwxr-xr-x 3 ubuntu ubuntu 76 Apr 1 14:30 /home/ubuntu/cephtest 2026-04-01T14:30:21.629 INFO:teuthology.orchestra.run.vm08.stdout: 58721128 0 drwxr-xr-x 2 ubuntu ubuntu 6 Apr 1 13:54 /home/ubuntu/cephtest/ceph.data 2026-04-01T14:30:21.629 INFO:teuthology.orchestra.run.vm08.stdout: 83890472 4 -rw-r--r-- 1 ceph root 20 Apr 1 13:55 /home/ubuntu/cephtest/url_file 2026-04-01T14:30:21.629 INFO:teuthology.orchestra.run.vm08.stdout: 83890473 0 srwxr-xr-x 1 root root 0 Apr 1 13:55 /home/ubuntu/cephtest/rgw.opslog.ceph.client.1.sock 2026-04-01T14:30:21.630 INFO:teuthology.orchestra.run.vm08.stderr:rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty 2026-04-01T14:30:21.644 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-01T14:30:21.644 ERROR:teuthology.run_tasks:Manager failed: internal.base Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/task/internal/__init__.py", line 48, in base yield File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/kshtsk/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_0d46c98b681455ef9cd34c2d3c6981da4e9f1aea/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/remote.py", line 596, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/kshtsk/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/kshtsk/teuthology/teuthology/task/internal/__init__.py", line 53, in base run.wait( File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/kshtsk/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest' 2026-04-01T14:30:21.644 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-04-01T14:30:21.647 DEBUG:teuthology.run_tasks:Exception was not quenched, exiting: CommandFailedError: Command failed on vm06 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' 2026-04-01T14:30:21.648 INFO:teuthology.run:Summary data: description: rgw/dedup/{beast bluestore-bitmap fixed-3-rgw ignore-pg-availability overrides supported-distros/{rocky_latest} tasks/{0-install test_dedup}} duration: 2471.015862941742 failure_reason: 'Command failed on vm06 with status 1: ''adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user rm --uid foo.client.0 --purge-data --cluster ceph''' flavor: default owner: supriti sentry_event: null status: fail success: false 2026-04-01T14:30:21.648 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-04-01T14:30:21.661 INFO:teuthology.orchestra.run.vm09.stdout: 83886251 0 drwxr-xr-x 3 ubuntu ubuntu 95 Apr 1 14:30 /home/ubuntu/cephtest 2026-04-01T14:30:21.661 INFO:teuthology.orchestra.run.vm09.stdout: 58721128 0 drwxr-xr-x 2 ubuntu ubuntu 6 Apr 1 13:54 /home/ubuntu/cephtest/ceph.data 2026-04-01T14:30:21.661 INFO:teuthology.orchestra.run.vm09.stdout: 83890462 4 -rw-r--r-- 1 ubuntu ubuntu 409 Apr 1 13:54 /home/ubuntu/cephtest/ceph.monmap 2026-04-01T14:30:21.661 INFO:teuthology.orchestra.run.vm09.stdout: 83890463 4 -rw-r--r-- 1 ceph root 20 Apr 1 13:55 /home/ubuntu/cephtest/url_file 2026-04-01T14:30:21.661 INFO:teuthology.orchestra.run.vm09.stdout: 83890464 0 srwxr-xr-x 1 root root 0 Apr 1 13:55 /home/ubuntu/cephtest/rgw.opslog.ceph.client.2.sock 2026-04-01T14:30:21.662 INFO:teuthology.orchestra.run.vm09.stderr:rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty 2026-04-01T14:30:21.672 INFO:teuthology.run:FAIL