2026-04-17T12:15:58.783 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-04-17T12:15:58.789 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-04-17T12:15:58.811 INFO:teuthology.run:Config: archive_path: /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589 branch: wip-sse-s3-on-v20.2.0 description: rgw/dedup/{beast bluestore-bitmap fixed-3-rgw ignore-pg-availability overrides supported-distros/{rocky_latest} tasks/{0-install test_dedup}} email: null first_in_suite: false flavor: default job_id: '5589' last_in_suite: false machine_type: vps name: supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: rocky os_version: '9.7' overrides: admin_socket: branch: wip-sse-s3-on-v20.2.0 ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: logical_volumes: lv_1: scratch_dev: true size: 25%VG vg: vg_nvme lv_2: scratch_dev: true size: 25%VG vg: vg_nvme lv_3: scratch_dev: true size: 25%VG vg: vg_nvme lv_4: scratch_dev: true size: 25%VG vg: vg_nvme timezone: UTC volume_groups: vg_nvme: pvs: /dev/vdb,/dev/vdc,/dev/vdd,/dev/vde ceph: conf: client: debug rgw: 20 debug rgw dedup: 20 setgroup: ceph setuser: ceph global: osd_max_pg_log_entries: 10 osd_min_pg_log_entries: 10 mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: bdev async discard: true bdev enable discard: true bluestore allocator: bitmap bluestore block size: 96636764160 bluestore fsck on mount: true debug bluefs: 1/20 debug bluestore: 1/20 debug ms: 1 debug osd: 20 debug rocksdb: 4/10 mon osd backfillfull_ratio: 0.85 mon osd full ratio: 0.9 mon osd nearfull ratio: 0.8 osd failsafe full ratio: 0.95 osd mclock iops capacity threshold hdd: 49000 osd objectstore: bluestore osd shutdown pgref assert: true flavor: default fs: xfs log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - \(PG_AVAILABILITY\) - \(PG_DEGRADED\) - \(POOL_APP_NOT_ENABLED\) - not have an application enabled sha1: c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8 ceph-deploy: bluestore: true conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} osd: bdev async discard: true bdev enable discard: true bluestore block size: 96636764160 bluestore fsck on mount: true debug bluefs: 1/20 debug bluestore: 1/20 debug rocksdb: 4/10 mon osd backfillfull_ratio: 0.85 mon osd full ratio: 0.9 mon osd nearfull ratio: 0.8 osd failsafe full ratio: 0.95 osd objectstore: bluestore fs: xfs cephadm: cephadm_binary_url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm containers: image: harbor.clyso.com/custom-ceph/ceph/ceph:sse-s3-kmip-preview-not-for-production-7 install: ceph: flavor: default sha1: c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8 extra_system_packages: deb: - python3-jmespath - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-jmespath - python3-xmltodict - s3cmd repos: - name: ceph-source priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS - name: ceph-noarch priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch - name: ceph priority: 1 url: https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64 rgw: frontend: beast storage classes: FROZEN: null LUKEWARM: null s3tests: sha1: e0c4ff71baef6d5126a0201df5fe54196d89b296 selinux: allowlist: - scontext=system_u:system_r:getty_t:s0 thrashosds: bdev_inject_crash: 2 bdev_inject_crash_probability: 0.5 workunit: branch: tt-wip-sse-s3-on-v20.2.0 sha1: 909b66e106532fd1f1a49171c3f2eb7a193a6d0b owner: supriti priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - mon.a - mon.c - mgr.y - osd.0 - osd.1 - osd.2 - osd.3 - client.0 - - mon.b - mgr.x - osd.4 - osd.5 - osd.6 - osd.7 - client.1 - - client.2 seed: 7685 sha1: c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8 sleep_before_teardown: 0 suite: rgw suite_branch: tt-wip-sse-s3-on-v20.2.0 suite_path: /home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa suite_relpath: qa suite_repo: http://git.local/ceph.git suite_sha1: 909b66e106532fd1f1a49171c3f2eb7a193a6d0b targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA3v6WWmK1n6Sfq3xIlK6cW2X2vPTSyNwHkSmjaFBSTp1tVknwZUk43AfMuvt2UVUlvhDs84IZdlA3iLH+lTTbw= vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDJh0VKrTxfmhFAF2A4azvEw3oyVDAGgxyD40VIJFd/pNOEXihGKEUMFa6EeCcdnot3hUQvEpW9Wu96wInC9a3E= vm08.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPtGs0s6hOkJKgF3AgFL0vCGkBCbq1q7LCFPDOhRpiT9seQ30oUK47pB0q4yI8pgBHC3VWBkTQLRETYTQROZIp4= tasks: - install: null - ceph: null - openssl_keys: null - rgw: - client.0 - client.1 - client.2 - tox: - client.0 - tox: - client.0 - dedup-tests: client.0: rgw_server: client.0 teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/kshtsk/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-04-17_12:11:56 tube: vps user: supriti verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.3072398 2026-04-17T12:15:58.811 INFO:teuthology.run:suite_path is set to /home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa; will attempt to use it 2026-04-17T12:15:58.812 INFO:teuthology.run:Found tasks at /home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks 2026-04-17T12:15:58.812 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-04-17T12:15:58.812 INFO:teuthology.task.internal:Saving configuration 2026-04-17T12:15:58.818 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-04-17T12:15:58.819 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-04-17T12:15:58.826 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'rocky', 'os_version': '9.7', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-04-17 12:14:31.104751', 'locked_by': 'supriti', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA3v6WWmK1n6Sfq3xIlK6cW2X2vPTSyNwHkSmjaFBSTp1tVknwZUk43AfMuvt2UVUlvhDs84IZdlA3iLH+lTTbw='} 2026-04-17T12:15:58.832 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'rocky', 'os_version': '9.7', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-04-17 12:14:31.105538', 'locked_by': 'supriti', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDJh0VKrTxfmhFAF2A4azvEw3oyVDAGgxyD40VIJFd/pNOEXihGKEUMFa6EeCcdnot3hUQvEpW9Wu96wInC9a3E='} 2026-04-17T12:15:58.838 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm08.local', 'description': '/archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'rocky', 'os_version': '9.7', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-04-17 12:14:31.105284', 'locked_by': 'supriti', 'mac_address': '52:55:00:00:00:08', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPtGs0s6hOkJKgF3AgFL0vCGkBCbq1q7LCFPDOhRpiT9seQ30oUK47pB0q4yI8pgBHC3VWBkTQLRETYTQROZIp4='} 2026-04-17T12:15:58.838 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-04-17T12:15:58.838 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['mon.a', 'mon.c', 'mgr.y', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'] 2026-04-17T12:15:58.839 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['mon.b', 'mgr.x', 'osd.4', 'osd.5', 'osd.6', 'osd.7', 'client.1'] 2026-04-17T12:15:58.839 INFO:teuthology.task.internal:roles: ubuntu@vm08.local - ['client.2'] 2026-04-17T12:15:58.839 INFO:teuthology.run_tasks:Running task console_log... 2026-04-17T12:15:58.846 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-04-17T12:15:58.852 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-04-17T12:15:58.858 DEBUG:teuthology.task.console_log:vm08 does not support IPMI; excluding 2026-04-17T12:15:58.858 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7f338dbe43a0>, signals=[15]) 2026-04-17T12:15:58.858 INFO:teuthology.run_tasks:Running task internal.connect... 2026-04-17T12:15:58.859 INFO:teuthology.task.internal:Opening connections... 2026-04-17T12:15:58.859 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-04-17T12:15:58.860 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-17T12:15:58.921 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-04-17T12:15:58.922 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-17T12:15:58.983 DEBUG:teuthology.task.internal:connecting to ubuntu@vm08.local 2026-04-17T12:15:58.984 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-17T12:15:59.044 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-04-17T12:15:59.045 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-04-17T12:15:59.062 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-04-17T12:15:59.062 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:NAME="Rocky Linux" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="9.7 (Blue Onyx)" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:ID="rocky" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE="rhel centos fedora" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="9.7" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:PLATFORM_ID="platform:el9" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="Rocky Linux 9.7 (Blue Onyx)" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:ANSI_COLOR="0;32" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:LOGO="fedora-logo-icon" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:CPE_NAME="cpe:/o:rocky:rocky:9::baseos" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://rockylinux.org/" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:VENDOR_NAME="RESF" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:VENDOR_URL="https://resf.org/" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://bugs.rockylinux.org/" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:SUPPORT_END="2032-05-31" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:ROCKY_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT="Rocky Linux" 2026-04-17T12:15:59.119 INFO:teuthology.orchestra.run.vm02.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-17T12:15:59.120 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-04-17T12:15:59.124 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-04-17T12:15:59.141 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-04-17T12:15:59.141 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:NAME="Rocky Linux" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9.7 (Blue Onyx)" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:ID="rocky" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel centos fedora" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9.7" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="Rocky Linux 9.7 (Blue Onyx)" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;32" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:rocky:rocky:9::baseos" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://rockylinux.org/" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:VENDOR_NAME="RESF" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:VENDOR_URL="https://resf.org/" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://bugs.rockylinux.org/" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:SUPPORT_END="2032-05-31" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:ROCKY_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Rocky Linux" 2026-04-17T12:15:59.197 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-17T12:15:59.197 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-04-17T12:15:59.203 DEBUG:teuthology.orchestra.run.vm08:> uname -m 2026-04-17T12:15:59.221 INFO:teuthology.orchestra.run.vm08.stdout:x86_64 2026-04-17T12:15:59.221 DEBUG:teuthology.orchestra.run.vm08:> cat /etc/os-release 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:NAME="Rocky Linux" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:VERSION="9.7 (Blue Onyx)" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:ID="rocky" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:ID_LIKE="rhel centos fedora" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:VERSION_ID="9.7" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:PLATFORM_ID="platform:el9" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:PRETTY_NAME="Rocky Linux 9.7 (Blue Onyx)" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:ANSI_COLOR="0;32" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:LOGO="fedora-logo-icon" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:CPE_NAME="cpe:/o:rocky:rocky:9::baseos" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:HOME_URL="https://rockylinux.org/" 2026-04-17T12:15:59.276 INFO:teuthology.orchestra.run.vm08.stdout:VENDOR_NAME="RESF" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:VENDOR_URL="https://resf.org/" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:BUG_REPORT_URL="https://bugs.rockylinux.org/" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:SUPPORT_END="2032-05-31" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:ROCKY_SUPPORT_PRODUCT="Rocky-Linux-9" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:ROCKY_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT="Rocky Linux" 2026-04-17T12:15:59.277 INFO:teuthology.orchestra.run.vm08.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="9.7" 2026-04-17T12:15:59.277 INFO:teuthology.lock.ops:Updating vm08.local on lock server 2026-04-17T12:15:59.282 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-04-17T12:15:59.284 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-04-17T12:15:59.285 INFO:teuthology.task.internal:Checking for old test directory... 2026-04-17T12:15:59.285 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-04-17T12:15:59.287 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-04-17T12:15:59.288 DEBUG:teuthology.orchestra.run.vm08:> test '!' -e /home/ubuntu/cephtest 2026-04-17T12:15:59.330 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-04-17T12:15:59.331 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-04-17T12:15:59.331 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-04-17T12:15:59.343 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-04-17T12:15:59.345 DEBUG:teuthology.orchestra.run.vm08:> test -z $(ls -A /var/lib/ceph) 2026-04-17T12:15:59.358 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-04-17T12:15:59.359 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-04-17T12:15:59.385 INFO:teuthology.orchestra.run.vm08.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-04-17T12:15:59.385 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-04-17T12:15:59.394 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-04-17T12:15:59.415 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:15:59.615 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-04-17T12:15:59.629 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:15:59.833 DEBUG:teuthology.orchestra.run.vm08:> test -e /ceph-qa-ready 2026-04-17T12:15:59.849 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:16:00.057 INFO:teuthology.run_tasks:Running task internal.base... 2026-04-17T12:16:00.059 INFO:teuthology.task.internal:Creating test directory... 2026-04-17T12:16:00.059 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-04-17T12:16:00.061 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-04-17T12:16:00.063 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-04-17T12:16:00.080 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-04-17T12:16:00.081 INFO:teuthology.run_tasks:Running task internal.archive... 2026-04-17T12:16:00.082 INFO:teuthology.task.internal:Creating archive directory... 2026-04-17T12:16:00.082 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-04-17T12:16:00.116 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-04-17T12:16:00.121 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-04-17T12:16:00.141 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-04-17T12:16:00.142 INFO:teuthology.task.internal:Enabling coredump saving... 2026-04-17T12:16:00.142 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-04-17T12:16:00.183 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:16:00.183 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-04-17T12:16:00.196 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:16:00.196 DEBUG:teuthology.orchestra.run.vm08:> test -f /run/.containerenv -o -f /.dockerenv 2026-04-17T12:16:00.213 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:16:00.213 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-04-17T12:16:00.225 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-04-17T12:16:00.238 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-04-17T12:16:00.248 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-17T12:16:00.256 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-17T12:16:00.260 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-17T12:16:00.269 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-17T12:16:00.281 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-17T12:16:00.292 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-04-17T12:16:00.293 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-04-17T12:16:00.294 INFO:teuthology.task.internal:Configuring sudo... 2026-04-17T12:16:00.295 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-04-17T12:16:00.299 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-04-17T12:16:00.313 DEBUG:teuthology.orchestra.run.vm08:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-04-17T12:16:00.363 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-04-17T12:16:00.365 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-04-17T12:16:00.365 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-04-17T12:16:00.367 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-04-17T12:16:00.378 DEBUG:teuthology.orchestra.run.vm08:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-04-17T12:16:00.420 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-04-17T12:16:00.448 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-04-17T12:16:00.510 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:16:00.510 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-04-17T12:16:00.569 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-04-17T12:16:00.594 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-04-17T12:16:00.648 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:16:00.648 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-04-17T12:16:00.710 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-04-17T12:16:00.739 DEBUG:teuthology.orchestra.run.vm08:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-04-17T12:16:00.798 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:16:00.799 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-04-17T12:16:00.860 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-04-17T12:16:00.861 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-04-17T12:16:00.863 DEBUG:teuthology.orchestra.run.vm08:> sudo service rsyslog restart 2026-04-17T12:16:00.889 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-17T12:16:00.897 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-17T12:16:00.935 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-17T12:16:01.284 INFO:teuthology.run_tasks:Running task internal.timer... 2026-04-17T12:16:01.285 INFO:teuthology.task.internal:Starting timer... 2026-04-17T12:16:01.285 INFO:teuthology.run_tasks:Running task pcp... 2026-04-17T12:16:01.288 INFO:teuthology.run_tasks:Running task selinux... 2026-04-17T12:16:01.290 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:getty_t:s0']} 2026-04-17T12:16:01.290 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-04-17T12:16:01.290 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-04-17T12:16:01.290 INFO:teuthology.task.selinux:Excluding vm08: VMs are not yet supported 2026-04-17T12:16:01.290 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-04-17T12:16:01.290 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-04-17T12:16:01.290 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-04-17T12:16:01.290 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-04-17T12:16:01.292 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}} 2026-04-17T12:16:01.292 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-04-17T12:16:01.293 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-04-17T12:16:01.948 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-04-17T12:16:01.955 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-04-17T12:16:01.955 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "logical_volumes": {"lv_1": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_2": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_3": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_4": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}}, "timezone": "UTC", "volume_groups": {"vg_nvme": {"pvs": "/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde"}}}' -i /tmp/teuth_ansible_inventory7lwxvq3m --limit vm02.local,vm06.local,vm08.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-04-17T12:18:02.518 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm08.local')] 2026-04-17T12:18:02.519 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-04-17T12:18:02.519 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-17T12:18:02.583 DEBUG:teuthology.orchestra.run.vm02:> true 2026-04-17T12:18:02.674 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-04-17T12:18:02.674 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-04-17T12:18:02.675 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-17T12:18:02.743 DEBUG:teuthology.orchestra.run.vm06:> true 2026-04-17T12:18:02.836 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-04-17T12:18:02.836 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm08.local' 2026-04-17T12:18:02.837 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm08.local', 'username': 'ubuntu', 'timeout': 60} 2026-04-17T12:18:02.904 DEBUG:teuthology.orchestra.run.vm08:> true 2026-04-17T12:18:02.997 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm08.local' 2026-04-17T12:18:02.997 INFO:teuthology.run_tasks:Running task clock... 2026-04-17T12:18:03.001 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-04-17T12:18:03.001 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-04-17T12:18:03.001 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-17T12:18:03.003 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-04-17T12:18:03.003 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-17T12:18:03.005 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-04-17T12:18:03.005 DEBUG:teuthology.orchestra.run.vm08:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-17T12:18:03.041 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-04-17T12:18:03.048 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-04-17T12:18:03.056 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-04-17T12:18:03.068 INFO:teuthology.orchestra.run.vm02.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-04-17T12:18:03.082 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-04-17T12:18:03.087 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-04-17T12:18:03.094 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-04-17T12:18:03.097 INFO:teuthology.orchestra.run.vm02.stderr:sudo: ntpd: command not found 2026-04-17T12:18:03.107 INFO:teuthology.orchestra.run.vm08.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-04-17T12:18:03.110 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-04-17T12:18:03.112 INFO:teuthology.orchestra.run.vm02.stdout:506 Cannot talk to daemon 2026-04-17T12:18:03.126 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-04-17T12:18:03.129 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-04-17T12:18:03.141 INFO:teuthology.orchestra.run.vm08.stderr:sudo: ntpd: command not found 2026-04-17T12:18:03.148 INFO:teuthology.orchestra.run.vm02.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-04-17T12:18:03.156 INFO:teuthology.orchestra.run.vm08.stdout:506 Cannot talk to daemon 2026-04-17T12:18:03.174 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-04-17T12:18:03.192 INFO:teuthology.orchestra.run.vm08.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-04-17T12:18:03.195 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-04-17T12:18:03.199 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-17T12:18:03.199 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-04-17T12:18:03.207 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-04-17T12:18:03.210 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-17T12:18:03.210 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-04-17T12:18:03.256 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-04-17T12:18:03.259 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-17T12:18:03.259 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-04-17T12:18:03.260 INFO:teuthology.run_tasks:Running task install... 2026-04-17T12:18:03.262 DEBUG:teuthology.task.install:project ceph 2026-04-17T12:18:03.262 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}]} 2026-04-17T12:18:03.263 DEBUG:teuthology.task.install:config {'flavor': 'default', 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8', 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-04-17T12:18:03.263 INFO:teuthology.task.install:Using flavor: default 2026-04-17T12:18:03.266 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-04-17T12:18:03.266 INFO:teuthology.task.install:extra packages: [] 2026-04-17T12:18:03.266 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}]} 2026-04-17T12:18:03.266 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}] 2026-04-17T12:18:03.266 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:18:03.266 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-04-17T12:18:03.266 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}]} 2026-04-17T12:18:03.266 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}] 2026-04-17T12:18:03.266 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:18:03.266 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-04-17T12:18:03.266 DEBUG:teuthology.task.install.rpm:_update_package_list_and_install: config is {'branch': None, 'cleanup': None, 'debuginfo': None, 'downgrade_packages': [], 'exclude_packages': [], 'extra_packages': [], 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'extras': None, 'enable_coprs': [], 'flavor': 'default', 'install_ceph_packages': True, 'packages': {}, 'project': 'ceph', 'repos_only': False, 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8', 'tag': None, 'wait_for_package': False, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}]} 2026-04-17T12:18:03.267 DEBUG:teuthology.task.install.rpm:Adding repos: [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}] 2026-04-17T12:18:03.267 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:18:03.267 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph-source.repo 2026-04-17T12:18:03.302 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:18:03.303 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-04-17T12:18:03.303 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:18:03.303 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-04-17T12:18:03.304 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:18:03.304 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph-noarch.repo 2026-04-17T12:18:03.378 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:18:03.378 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-04-17T12:18:03.385 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:18:03.385 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-04-17T12:18:03.388 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:18:03.388 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/yum.repos.d/ceph.repo 2026-04-17T12:18:03.454 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-jmespath, python3-xmltodict, s3cmd on remote rpm x86_64 2026-04-17T12:18:03.454 DEBUG:teuthology.orchestra.run.vm02:> sudo yum clean all 2026-04-17T12:18:03.459 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-jmespath, python3-xmltodict, s3cmd on remote rpm x86_64 2026-04-17T12:18:03.459 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-04-17T12:18:03.464 INFO:teuthology.task.install.rpm:Installing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd, bzip2, perl-Test-Harness, python3-jmespath, python3-xmltodict, s3cmd on remote rpm x86_64 2026-04-17T12:18:03.464 DEBUG:teuthology.orchestra.run.vm08:> sudo yum clean all 2026-04-17T12:18:03.665 INFO:teuthology.orchestra.run.vm06.stdout:47 files removed 2026-04-17T12:18:03.666 INFO:teuthology.orchestra.run.vm08.stdout:47 files removed 2026-04-17T12:18:03.667 INFO:teuthology.orchestra.run.vm02.stdout:47 files removed 2026-04-17T12:18:03.693 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-jmespath python3-xmltodict s3cmd 2026-04-17T12:18:03.709 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-jmespath python3-xmltodict s3cmd 2026-04-17T12:18:03.711 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y install ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd bzip2 perl-Test-Harness python3-jmespath python3-xmltodict s3cmd 2026-04-17T12:18:04.285 INFO:teuthology.orchestra.run.vm08.stdout:ceph 258 kB/s | 90 kB 00:00 2026-04-17T12:18:04.285 INFO:teuthology.orchestra.run.vm02.stdout:ceph 254 kB/s | 90 kB 00:00 2026-04-17T12:18:04.287 INFO:teuthology.orchestra.run.vm06.stdout:ceph 242 kB/s | 90 kB 00:00 2026-04-17T12:18:04.557 INFO:teuthology.orchestra.run.vm02.stdout:ceph-noarch 103 kB/s | 25 kB 00:00 2026-04-17T12:18:04.559 INFO:teuthology.orchestra.run.vm08.stdout:ceph-noarch 101 kB/s | 25 kB 00:00 2026-04-17T12:18:04.566 INFO:teuthology.orchestra.run.vm06.stdout:ceph-noarch 101 kB/s | 25 kB 00:00 2026-04-17T12:18:04.757 INFO:teuthology.orchestra.run.vm02.stdout:ceph-source 13 kB/s | 2.3 kB 00:00 2026-04-17T12:18:04.758 INFO:teuthology.orchestra.run.vm08.stdout:ceph-source 13 kB/s | 2.3 kB 00:00 2026-04-17T12:18:04.761 INFO:teuthology.orchestra.run.vm06.stdout:ceph-source 13 kB/s | 2.3 kB 00:00 2026-04-17T12:18:05.267 INFO:teuthology.orchestra.run.vm08.stdout:Extra Packages for Enterprise Linux 42 MB/s | 21 MB 00:00 2026-04-17T12:18:05.424 INFO:teuthology.orchestra.run.vm02.stdout:Extra Packages for Enterprise Linux 32 MB/s | 21 MB 00:00 2026-04-17T12:18:05.542 INFO:teuthology.orchestra.run.vm06.stdout:Extra Packages for Enterprise Linux 27 MB/s | 21 MB 00:00 2026-04-17T12:18:10.360 INFO:teuthology.orchestra.run.vm08.stdout:lab-extras 65 kB/s | 50 kB 00:00 2026-04-17T12:18:10.510 INFO:teuthology.orchestra.run.vm02.stdout:lab-extras 63 kB/s | 50 kB 00:00 2026-04-17T12:18:10.523 INFO:teuthology.orchestra.run.vm06.stdout:lab-extras 63 kB/s | 50 kB 00:00 2026-04-17T12:18:11.776 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - BaseOS 15 MB/s | 19 MB 00:01 2026-04-17T12:18:11.888 INFO:teuthology.orchestra.run.vm02.stdout:Rocky Linux 9 - BaseOS 15 MB/s | 19 MB 00:01 2026-04-17T12:18:12.085 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - BaseOS 13 MB/s | 19 MB 00:01 2026-04-17T12:18:13.827 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - AppStream 22 MB/s | 18 MB 00:00 2026-04-17T12:18:13.998 INFO:teuthology.orchestra.run.vm02.stdout:Rocky Linux 9 - AppStream 27 MB/s | 18 MB 00:00 2026-04-17T12:18:14.247 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - AppStream 27 MB/s | 18 MB 00:00 2026-04-17T12:18:16.359 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - CRB 8.4 MB/s | 4.5 MB 00:00 2026-04-17T12:18:16.568 INFO:teuthology.orchestra.run.vm02.stdout:Rocky Linux 9 - CRB 8.4 MB/s | 4.5 MB 00:00 2026-04-17T12:18:16.788 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - CRB 8.2 MB/s | 4.5 MB 00:00 2026-04-17T12:18:17.350 INFO:teuthology.orchestra.run.vm08.stdout:Rocky Linux 9 - Extras 49 kB/s | 17 kB 00:00 2026-04-17T12:18:17.558 INFO:teuthology.orchestra.run.vm02.stdout:Rocky Linux 9 - Extras 48 kB/s | 17 kB 00:00 2026-04-17T12:18:17.749 INFO:teuthology.orchestra.run.vm06.stdout:Rocky Linux 9 - Extras 49 kB/s | 17 kB 00:00 2026-04-17T12:18:18.620 INFO:teuthology.orchestra.run.vm08.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-17T12:18:18.621 INFO:teuthology.orchestra.run.vm08.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-17T12:18:18.651 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================== 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================== 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout:Installing: 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: bzip2 x86_64 1.0.8-10.el9_5 baseos 51 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 6.5 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 5.9 M 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 940 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 154 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 962 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 173 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 15 M 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 7.4 M 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 50 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 M 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 85 M 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 297 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 1.0 M 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 34 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 868 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 126 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 267 k 2026-04-17T12:18:18.657 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 163 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath noarch 1.0.1-1.el9_7 appstream 43 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 317 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 305 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 99 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 91 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.9 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 180 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout:Upgrading: 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: librados2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 3.5 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: librbd1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.8 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout:Installing dependencies: 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options x86_64 1.75.0-13.el9_7 appstream 104 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: c-ares x86_64 1.19.1-2.el9_4 baseos 110 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 43 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.3 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 289 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 5.0 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 17 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 17 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 25 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.7.2-4.el9 baseos 310 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 appstream 30 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 appstream 3.0 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 appstream 15 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: fuse x86_64 2.9.9-17.el9 baseos 78 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 41 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-proxy2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 164 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 baseos 71 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-11.el9 baseos 794 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: libnbd x86_64 1.20.3-4.el9 appstream 171 k 2026-04-17T12:18:18.658 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 159 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-11.el9 baseos 184 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 44 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 250 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: librgw2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 6.4 M 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 243 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-14.el9_7.1 appstream 240 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 282 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 appstream 187 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 crb 21 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9.0.2 baseos 32 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 appstream 41 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 baseos 92 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: perl-Benchmark noarch 1.23-481.1.el9_6 appstream 25 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9_7 appstream 1.0 M 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 crb 862 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 appstream 5.8 M 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 45 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 163 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 241 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-5.el9 epel 173 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.10.0-5.el9 epel 290 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 baseos 1.2 M 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.23-2.el9 appstream 205 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-04-17T12:18:18.659 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 appstream 228 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 166 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 32 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 appstream 5.8 M 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 appstream 368 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 appstream 69 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9.0.1 baseos 103 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9_7 appstream 237 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 appstream 132 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 appstream 210 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 baseos 124 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 baseos 150 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9_6 baseos 115 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 43 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9.0.1 appstream 44 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 baseos 191 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 24.09.0-1.el9 appstream 221 k 2026-04-17T12:18:18.660 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 65 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 appstream 299 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 baseos 180 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 188 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 89 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 63 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 baseos 263 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout:Installing weak dependencies: 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 22 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-daemon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 35 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli x86_64 2.13-1.el9 baseos 1.0 M 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 24.09.0-1.el9 appstream 36 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools x86_64 1:7.2-9.el9 baseos 551 k 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout:============================================================================================== 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout:Install 149 Packages 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout:Upgrade 2 Packages 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:18.661 INFO:teuthology.orchestra.run.vm08.stdout:Total download size: 274 M 2026-04-17T12:18:18.662 INFO:teuthology.orchestra.run.vm08.stdout:Downloading Packages: 2026-04-17T12:18:18.778 INFO:teuthology.orchestra.run.vm02.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-17T12:18:18.778 INFO:teuthology.orchestra.run.vm02.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-17T12:18:18.802 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================================== 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================================== 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout:Installing: 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: bzip2 x86_64 1.0.8-10.el9_5 baseos 51 k 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 6.5 k 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 5.9 M 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 940 k 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 154 k 2026-04-17T12:18:18.807 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 962 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 173 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 15 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 7.4 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 50 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 85 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 297 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 1.0 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 34 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 868 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 126 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 267 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 163 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath noarch 1.0.1-1.el9_7 appstream 43 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 317 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 305 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 99 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 91 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.9 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 180 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout:Upgrading: 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: librados2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 3.5 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: librbd1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.8 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout:Installing dependencies: 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options x86_64 1.75.0-13.el9_7 appstream 104 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: c-ares x86_64 1.19.1-2.el9_4 baseos 110 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 43 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.3 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 289 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 5.0 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 17 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 17 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 25 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.7.2-4.el9 baseos 310 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 appstream 30 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 appstream 3.0 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 appstream 15 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: fuse x86_64 2.9.9-17.el9 baseos 78 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 41 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-04-17T12:18:18.808 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 164 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 baseos 71 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-11.el9 baseos 794 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libnbd x86_64 1.20.3-4.el9 appstream 171 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 159 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-11.el9 baseos 184 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 44 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 250 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: librgw2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 6.4 M 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 243 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-14.el9_7.1 appstream 240 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 282 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 appstream 187 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 crb 21 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9.0.2 baseos 32 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 appstream 41 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 baseos 92 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: perl-Benchmark noarch 1.23-481.1.el9_6 appstream 25 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9_7 appstream 1.0 M 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 crb 862 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 appstream 5.8 M 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 45 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 163 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 241 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-5.el9 epel 173 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.10.0-5.el9 epel 290 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 baseos 1.2 M 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.23-2.el9 appstream 205 k 2026-04-17T12:18:18.809 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 appstream 228 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 166 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 32 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 appstream 5.8 M 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 appstream 368 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 appstream 69 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9.0.1 baseos 103 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9_7 appstream 237 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 appstream 132 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 appstream 210 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 baseos 124 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 baseos 150 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9_6 baseos 115 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 43 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9.0.1 appstream 44 k 2026-04-17T12:18:18.810 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 baseos 191 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 24.09.0-1.el9 appstream 221 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 65 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 appstream 299 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 baseos 180 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 188 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 89 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 63 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 baseos 263 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:Installing weak dependencies: 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 22 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 35 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli x86_64 2.13-1.el9 baseos 1.0 M 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 24.09.0-1.el9 appstream 36 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools x86_64 1:7.2-9.el9 baseos 551 k 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================================== 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:Install 149 Packages 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:Upgrade 2 Packages 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:Total download size: 274 M 2026-04-17T12:18:18.811 INFO:teuthology.orchestra.run.vm02.stdout:Downloading Packages: 2026-04-17T12:18:18.998 INFO:teuthology.orchestra.run.vm06.stdout:Package librados2-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-17T12:18:18.998 INFO:teuthology.orchestra.run.vm06.stdout:Package librbd1-2:16.2.4-5.el9.x86_64 is already installed. 2026-04-17T12:18:19.024 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================== 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================== 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: bzip2 x86_64 1.0.8-10.el9_5 baseos 51 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 6.5 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 5.9 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 940 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 154 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 962 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 173 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 15 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 7.4 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 50 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 85 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 297 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 1.0 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 34 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 868 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 126 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: perl-Test-Harness noarch 1:3.42-461.el9 appstream 267 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 163 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9_7 appstream 43 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 317 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 305 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 99 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict noarch 0.12.0-15.el9 epel 22 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 91 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.9 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 180 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: s3cmd noarch 2.4.0-1.el9 epel 206 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout:Upgrading: 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: librados2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 3.5 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: librbd1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.8 M 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp x86_64 20211102.0-4.el9 epel 551 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options x86_64 1.75.0-13.el9_7 appstream 104 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: c-ares x86_64 1.19.1-2.el9_4 baseos 110 k 2026-04-17T12:18:19.029 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 43 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 2.3 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 289 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 5.0 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 17 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 17 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 25 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup x86_64 2.7.2-4.el9 baseos 310 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 appstream 30 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 appstream 3.0 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 appstream 15 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: fuse x86_64 2.9.9-17.el9 baseos 78 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs x86_64 2.9.1-3.el9 epel 308 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data noarch 1.46.7-10.el9 epel 19 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs x86_64 1.1.0-3.el9 baseos 41 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libarrow x86_64 9.0.0-15.el9 epel 4.4 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc noarch 9.0.0-15.el9 epel 25 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-proxy2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 24 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 164 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libconfig x86_64 1.7.2-9.el9 baseos 71 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran x86_64 11.5.0-11.el9 baseos 794 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libnbd x86_64 1.20.3-4.el9 appstream 171 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: liboath x86_64 2.6.12-1.el9 epel 49 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj x86_64 1.12.1-1.el9 appstream 159 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath x86_64 11.5.0-11.el9 baseos 184 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq x86_64 0.11.0-7.el9 appstream 44 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 250 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka x86_64 1.6.1-102.el9 appstream 662 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: librgw2 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 6.4 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt x86_64 1.10.1-1.el9 appstream 243 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libunwind x86_64 1.6.2-1.el9 epel 67 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: libxslt x86_64 1.1.34-14.el9_7.1 appstream 240 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust x86_64 2.12.0-6.el9 appstream 282 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: lua x86_64 5.4.4-4.el9 appstream 187 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel x86_64 5.4.4-4.el9 crb 21 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: luarocks noarch 3.9.2-5.el9 epel 151 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: mailcap noarch 2.1.49-5.el9.0.2 baseos 32 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: openblas x86_64 0.3.29-1.el9 appstream 41 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp x86_64 0.3.29-1.el9 appstream 5.3 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs x86_64 9.0.0-15.el9 epel 838 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: pciutils x86_64 3.7.0-7.el9 baseos 92 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: perl-Benchmark noarch 1.23-481.1.el9_6 appstream 25 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: protobuf x86_64 3.14.0-17.el9_7 appstream 1.0 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 crb 862 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh noarch 2.13.2-5.el9 epel 548 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand noarch 2.2.2-8.el9 epel 29 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel noarch 2.9.1-2.el9 appstream 5.8 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 epel 60 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt x86_64 3.2.2-1.el9 epel 43 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools noarch 4.2.4-1.el9 epel 32 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 45 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 163 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi noarch 2023.05.07-4.el9 epel 14 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi x86_64 1.14.5-5.el9 baseos 241 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot noarch 10.0.1-5.el9 epel 173 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy noarch 18.10.0-5.el9 epel 290 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 baseos 1.2 M 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel x86_64 3.9.23-2.el9 appstream 205 k 2026-04-17T12:18:19.030 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth noarch 1:2.45.0-1.el9 epel 254 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio x86_64 1.46.7-10.el9 epel 2.0 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 epel 144 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate noarch 0.6.1-3.el9 epel 56 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco noarch 8.2.1-3.el9 epel 11 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 epel 18 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 epel 23 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context noarch 6.0.1-3.el9 epel 20 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 epel 19 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text noarch 4.0.0-2.el9 epel 26 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 appstream 228 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 epel 1.0 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 appstream 166 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml x86_64 4.6.5-3.el9 appstream 1.2 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe x86_64 1.1.1-12.el9 appstream 32 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools noarch 8.12.0-2.el9 epel 79 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack x86_64 1.0.3-2.el9 epel 86 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort noarch 7.1.1-5.el9 epel 58 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 appstream 5.8 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 appstream 368 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging noarch 20.9-5.el9 appstream 69 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply noarch 3.11-14.el9.0.1 baseos 103 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend noarch 3.1.0-2.el9 epel 16 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf noarch 3.14.0-17.el9_7 appstream 237 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 epel 90 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 appstream 132 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 appstream 210 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser noarch 2.20-6.el9 baseos 124 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 baseos 150 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru noarch 0.7-16.el9 epel 31 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests noarch 2.25.1-10.el9_6 baseos 115 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 appstream 43 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes noarch 2.5.1-5.el9 epel 188 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa noarch 4.9-2.el9 epel 59 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy x86_64 1.9.3-2.el9 appstream 19 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora noarch 5.0.0-2.el9 epel 36 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml noarch 0.10.2-6.el9.0.1 appstream 44 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions noarch 4.15.0-1.el9 epel 86 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 baseos 191 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client noarch 1.2.3-2.el9 epel 90 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec x86_64 1.3.13-1.el9 epel 48 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile noarch 2.0-10.el9 epel 20 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: qatlib x86_64 24.09.0-1.el9 appstream 221 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs x86_64 1.3.1-1.el9 appstream 65 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: re2 x86_64 1:20211101-20.el9 epel 191 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: socat x86_64 1.7.4.1-8.el9 appstream 299 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: thrift x86_64 0.15.0-4.el9 epel 1.6 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: unzip x86_64 6.0-59.el9 baseos 180 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1 x86_64 1.2.29-13.el9 appstream 188 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 appstream 89 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet x86_64 1.6.1-20.el9 appstream 63 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: zip x86_64 3.0-35.el9 baseos 263 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout:Installing weak dependencies: 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph-noarch 22 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso ceph 35 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.13-1.el9 baseos 1.0 M 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb noarch 5.3.1-1.el9 epel 139 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml noarch 1.16.0-1.el9 epel 125 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service x86_64 24.09.0-1.el9 appstream 36 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools x86_64 1:7.2-9.el9 baseos 551 k 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout:============================================================================================== 2026-04-17T12:18:19.031 INFO:teuthology.orchestra.run.vm06.stdout:Install 149 Packages 2026-04-17T12:18:19.032 INFO:teuthology.orchestra.run.vm06.stdout:Upgrade 2 Packages 2026-04-17T12:18:19.032 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:19.032 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 274 M 2026-04-17T12:18:19.032 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-04-17T12:18:19.502 INFO:teuthology.orchestra.run.vm02.stdout:(1/151): ceph-20.2.0-21.gc03ba9ecf58.el9.clyso. 69 kB/s | 6.5 kB 00:00 2026-04-17T12:18:19.710 INFO:teuthology.orchestra.run.vm06.stdout:(1/151): ceph-20.2.0-21.gc03ba9ecf58.el9.clyso. 78 kB/s | 6.5 kB 00:00 2026-04-17T12:18:19.769 INFO:teuthology.orchestra.run.vm02.stdout:(2/151): ceph-fuse-20.2.0-21.gc03ba9ecf58.el9.c 3.4 MB/s | 940 kB 00:00 2026-04-17T12:18:19.786 INFO:teuthology.orchestra.run.vm08.stdout:(1/151): ceph-20.2.0-21.gc03ba9ecf58.el9.clyso. 57 kB/s | 6.5 kB 00:00 2026-04-17T12:18:19.843 INFO:teuthology.orchestra.run.vm02.stdout:(3/151): ceph-immutable-object-cache-20.2.0-21. 2.1 MB/s | 154 kB 00:00 2026-04-17T12:18:19.959 INFO:teuthology.orchestra.run.vm06.stdout:(2/151): ceph-fuse-20.2.0-21.gc03ba9ecf58.el9.c 3.7 MB/s | 940 kB 00:00 2026-04-17T12:18:20.026 INFO:teuthology.orchestra.run.vm06.stdout:(3/151): ceph-immutable-object-cache-20.2.0-21. 2.2 MB/s | 154 kB 00:00 2026-04-17T12:18:20.204 INFO:teuthology.orchestra.run.vm08.stdout:(2/151): ceph-fuse-20.2.0-21.gc03ba9ecf58.el9.c 2.2 MB/s | 940 kB 00:00 2026-04-17T12:18:20.241 INFO:teuthology.orchestra.run.vm06.stdout:(4/151): ceph-mds-20.2.0-21.gc03ba9ecf58.el9.cl 11 MB/s | 2.3 MB 00:00 2026-04-17T12:18:20.253 INFO:teuthology.orchestra.run.vm08.stdout:(3/151): ceph-immutable-object-cache-20.2.0-21. 3.1 MB/s | 154 kB 00:00 2026-04-17T12:18:20.307 INFO:teuthology.orchestra.run.vm02.stdout:(4/151): ceph-common-20.2.0-21.gc03ba9ecf58.el9 26 MB/s | 24 MB 00:00 2026-04-17T12:18:20.450 INFO:teuthology.orchestra.run.vm06.stdout:(5/151): ceph-mgr-20.2.0-21.gc03ba9ecf58.el9.cl 4.5 MB/s | 962 kB 00:00 2026-04-17T12:18:20.535 INFO:teuthology.orchestra.run.vm02.stdout:(5/151): ceph-mgr-20.2.0-21.gc03ba9ecf58.el9.cl 4.1 MB/s | 962 kB 00:00 2026-04-17T12:18:20.554 INFO:teuthology.orchestra.run.vm06.stdout:(6/151): ceph-base-20.2.0-21.gc03ba9ecf58.el9.c 6.3 MB/s | 5.9 MB 00:00 2026-04-17T12:18:20.674 INFO:teuthology.orchestra.run.vm02.stdout:(6/151): ceph-mds-20.2.0-21.gc03ba9ecf58.el9.cl 2.8 MB/s | 2.3 MB 00:00 2026-04-17T12:18:20.817 INFO:teuthology.orchestra.run.vm06.stdout:(7/151): ceph-mon-20.2.0-21.gc03ba9ecf58.el9.cl 14 MB/s | 5.0 MB 00:00 2026-04-17T12:18:20.852 INFO:teuthology.orchestra.run.vm08.stdout:(4/151): ceph-mds-20.2.0-21.gc03ba9ecf58.el9.cl 3.9 MB/s | 2.3 MB 00:00 2026-04-17T12:18:20.996 INFO:teuthology.orchestra.run.vm02.stdout:(7/151): ceph-mon-20.2.0-21.gc03ba9ecf58.el9.cl 11 MB/s | 5.0 MB 00:00 2026-04-17T12:18:21.170 INFO:teuthology.orchestra.run.vm08.stdout:(5/151): ceph-mgr-20.2.0-21.gc03ba9ecf58.el9.cl 3.0 MB/s | 962 kB 00:00 2026-04-17T12:18:21.952 INFO:teuthology.orchestra.run.vm08.stdout:(6/151): ceph-base-20.2.0-21.gc03ba9ecf58.el9.c 2.6 MB/s | 5.9 MB 00:02 2026-04-17T12:18:22.043 INFO:teuthology.orchestra.run.vm08.stdout:(7/151): ceph-mon-20.2.0-21.gc03ba9ecf58.el9.cl 5.8 MB/s | 5.0 MB 00:00 2026-04-17T12:18:22.081 INFO:teuthology.orchestra.run.vm02.stdout:(8/151): ceph-osd-20.2.0-21.gc03ba9ecf58.el9.cl 12 MB/s | 17 MB 00:01 2026-04-17T12:18:22.136 INFO:teuthology.orchestra.run.vm02.stdout:(9/151): ceph-selinux-20.2.0-21.gc03ba9ecf58.el 460 kB/s | 25 kB 00:00 2026-04-17T12:18:22.260 INFO:teuthology.orchestra.run.vm06.stdout:(8/151): ceph-osd-20.2.0-21.gc03ba9ecf58.el9.cl 9.9 MB/s | 17 MB 00:01 2026-04-17T12:18:22.320 INFO:teuthology.orchestra.run.vm06.stdout:(9/151): ceph-selinux-20.2.0-21.gc03ba9ecf58.el 424 kB/s | 25 kB 00:00 2026-04-17T12:18:22.504 INFO:teuthology.orchestra.run.vm02.stdout:(10/151): ceph-radosgw-20.2.0-21.gc03ba9ecf58.e 16 MB/s | 24 MB 00:01 2026-04-17T12:18:22.545 INFO:teuthology.orchestra.run.vm02.stdout:(11/151): libcephfs-daemon-20.2.0-21.gc03ba9ecf 888 kB/s | 35 kB 00:00 2026-04-17T12:18:22.569 INFO:teuthology.orchestra.run.vm06.stdout:(10/151): ceph-radosgw-20.2.0-21.gc03ba9ecf58.e 14 MB/s | 24 MB 00:01 2026-04-17T12:18:22.601 INFO:teuthology.orchestra.run.vm02.stdout:(12/151): libcephfs-devel-20.2.0-21.gc03ba9ecf5 615 kB/s | 34 kB 00:00 2026-04-17T12:18:22.608 INFO:teuthology.orchestra.run.vm06.stdout:(11/151): libcephfs-daemon-20.2.0-21.gc03ba9ecf 913 kB/s | 35 kB 00:00 2026-04-17T12:18:22.641 INFO:teuthology.orchestra.run.vm02.stdout:(13/151): libcephfs-proxy2-20.2.0-21.gc03ba9ecf 607 kB/s | 24 kB 00:00 2026-04-17T12:18:22.648 INFO:teuthology.orchestra.run.vm06.stdout:(12/151): libcephfs-devel-20.2.0-21.gc03ba9ecf5 857 kB/s | 34 kB 00:00 2026-04-17T12:18:22.691 INFO:teuthology.orchestra.run.vm06.stdout:(13/151): libcephfs-proxy2-20.2.0-21.gc03ba9ecf 572 kB/s | 24 kB 00:00 2026-04-17T12:18:22.797 INFO:teuthology.orchestra.run.vm02.stdout:(14/151): ceph-base-20.2.0-21.gc03ba9ecf58.el9. 1.7 MB/s | 5.9 MB 00:03 2026-04-17T12:18:22.820 INFO:teuthology.orchestra.run.vm02.stdout:(15/151): libcephfs2-20.2.0-21.gc03ba9ecf58.el9 4.7 MB/s | 868 kB 00:00 2026-04-17T12:18:22.895 INFO:teuthology.orchestra.run.vm02.stdout:(16/151): libcephsqlite-20.2.0-21.gc03ba9ecf58. 1.6 MB/s | 164 kB 00:00 2026-04-17T12:18:22.925 INFO:teuthology.orchestra.run.vm02.stdout:(17/151): librados-devel-20.2.0-21.gc03ba9ecf58 1.2 MB/s | 126 kB 00:00 2026-04-17T12:18:22.943 INFO:teuthology.orchestra.run.vm06.stdout:(14/151): libcephfs2-20.2.0-21.gc03ba9ecf58.el9 3.4 MB/s | 868 kB 00:00 2026-04-17T12:18:23.002 INFO:teuthology.orchestra.run.vm06.stdout:(15/151): libcephsqlite-20.2.0-21.gc03ba9ecf58. 2.7 MB/s | 164 kB 00:00 2026-04-17T12:18:23.028 INFO:teuthology.orchestra.run.vm02.stdout:(18/151): libradosstriper1-20.2.0-21.gc03ba9ecf 1.9 MB/s | 250 kB 00:00 2026-04-17T12:18:23.088 INFO:teuthology.orchestra.run.vm02.stdout:(19/151): python3-ceph-argparse-20.2.0-21.gc03b 753 kB/s | 45 kB 00:00 2026-04-17T12:18:23.098 INFO:teuthology.orchestra.run.vm06.stdout:(16/151): librados-devel-20.2.0-21.gc03ba9ecf58 1.3 MB/s | 126 kB 00:00 2026-04-17T12:18:23.171 INFO:teuthology.orchestra.run.vm06.stdout:(17/151): libradosstriper1-20.2.0-21.gc03ba9ecf 3.4 MB/s | 250 kB 00:00 2026-04-17T12:18:23.185 INFO:teuthology.orchestra.run.vm02.stdout:(20/151): python3-ceph-common-20.2.0-21.gc03ba9 1.6 MB/s | 163 kB 00:00 2026-04-17T12:18:23.290 INFO:teuthology.orchestra.run.vm02.stdout:(21/151): python3-cephfs-20.2.0-21.gc03ba9ecf58 1.5 MB/s | 163 kB 00:00 2026-04-17T12:18:23.362 INFO:teuthology.orchestra.run.vm08.stdout:(8/151): ceph-common-20.2.0-21.gc03ba9ecf58.el9 6.5 MB/s | 24 MB 00:03 2026-04-17T12:18:23.399 INFO:teuthology.orchestra.run.vm08.stdout:(9/151): ceph-selinux-20.2.0-21.gc03ba9ecf58.el 686 kB/s | 25 kB 00:00 2026-04-17T12:18:23.519 INFO:teuthology.orchestra.run.vm02.stdout:(22/151): librgw2-20.2.0-21.gc03ba9ecf58.el9.cl 11 MB/s | 6.4 MB 00:00 2026-04-17T12:18:23.523 INFO:teuthology.orchestra.run.vm02.stdout:(23/151): python3-rados-20.2.0-21.gc03ba9ecf58. 1.3 MB/s | 317 kB 00:00 2026-04-17T12:18:23.583 INFO:teuthology.orchestra.run.vm02.stdout:(24/151): python3-rbd-20.2.0-21.gc03ba9ecf58.el 4.7 MB/s | 305 kB 00:00 2026-04-17T12:18:23.585 INFO:teuthology.orchestra.run.vm02.stdout:(25/151): python3-rgw-20.2.0-21.gc03ba9ecf58.el 1.6 MB/s | 99 kB 00:00 2026-04-17T12:18:23.630 INFO:teuthology.orchestra.run.vm02.stdout:(26/151): rbd-fuse-20.2.0-21.gc03ba9ecf58.el9.c 1.9 MB/s | 91 kB 00:00 2026-04-17T12:18:23.691 INFO:teuthology.orchestra.run.vm02.stdout:(27/151): rbd-nbd-20.2.0-21.gc03ba9ecf58.el9.cl 2.9 MB/s | 180 kB 00:00 2026-04-17T12:18:23.741 INFO:teuthology.orchestra.run.vm02.stdout:(28/151): ceph-grafana-dashboards-20.2.0-21.gc0 872 kB/s | 43 kB 00:00 2026-04-17T12:18:23.800 INFO:teuthology.orchestra.run.vm02.stdout:(29/151): ceph-mgr-cephadm-20.2.0-21.gc03ba9ecf 2.9 MB/s | 173 kB 00:00 2026-04-17T12:18:24.005 INFO:teuthology.orchestra.run.vm06.stdout:(18/151): librgw2-20.2.0-21.gc03ba9ecf58.el9.cl 7.6 MB/s | 6.4 MB 00:00 2026-04-17T12:18:24.038 INFO:teuthology.orchestra.run.vm06.stdout:(19/151): python3-ceph-argparse-20.2.0-21.gc03b 1.3 MB/s | 45 kB 00:00 2026-04-17T12:18:24.075 INFO:teuthology.orchestra.run.vm06.stdout:(20/151): python3-ceph-common-20.2.0-21.gc03ba9 4.4 MB/s | 163 kB 00:00 2026-04-17T12:18:24.112 INFO:teuthology.orchestra.run.vm06.stdout:(21/151): python3-cephfs-20.2.0-21.gc03ba9ecf58 4.3 MB/s | 163 kB 00:00 2026-04-17T12:18:24.191 INFO:teuthology.orchestra.run.vm06.stdout:(22/151): python3-rados-20.2.0-21.gc03ba9ecf58. 3.9 MB/s | 317 kB 00:00 2026-04-17T12:18:24.264 INFO:teuthology.orchestra.run.vm06.stdout:(23/151): python3-rbd-20.2.0-21.gc03ba9ecf58.el 4.1 MB/s | 305 kB 00:00 2026-04-17T12:18:24.304 INFO:teuthology.orchestra.run.vm06.stdout:(24/151): python3-rgw-20.2.0-21.gc03ba9ecf58.el 2.5 MB/s | 99 kB 00:00 2026-04-17T12:18:24.335 INFO:teuthology.orchestra.run.vm06.stdout:(25/151): rbd-fuse-20.2.0-21.gc03ba9ecf58.el9.c 2.9 MB/s | 91 kB 00:00 2026-04-17T12:18:24.545 INFO:teuthology.orchestra.run.vm02.stdout:(30/151): ceph-mgr-dashboard-20.2.0-21.gc03ba9e 20 MB/s | 15 MB 00:00 2026-04-17T12:18:24.546 INFO:teuthology.orchestra.run.vm06.stdout:(26/151): rbd-mirror-20.2.0-21.gc03ba9ecf58.el9 14 MB/s | 2.9 MB 00:00 2026-04-17T12:18:24.584 INFO:teuthology.orchestra.run.vm06.stdout:(27/151): rbd-nbd-20.2.0-21.gc03ba9ecf58.el9.cl 4.7 MB/s | 180 kB 00:00 2026-04-17T12:18:24.617 INFO:teuthology.orchestra.run.vm06.stdout:(28/151): ceph-grafana-dashboards-20.2.0-21.gc0 1.3 MB/s | 43 kB 00:00 2026-04-17T12:18:24.682 INFO:teuthology.orchestra.run.vm06.stdout:(29/151): ceph-mgr-cephadm-20.2.0-21.gc03ba9ecf 2.6 MB/s | 173 kB 00:00 2026-04-17T12:18:24.713 INFO:teuthology.orchestra.run.vm02.stdout:(31/151): rbd-mirror-20.2.0-21.gc03ba9ecf58.el9 2.6 MB/s | 2.9 MB 00:01 2026-04-17T12:18:24.788 INFO:teuthology.orchestra.run.vm02.stdout:(32/151): ceph-mgr-k8sevents-20.2.0-21.gc03ba9e 299 kB/s | 22 kB 00:00 2026-04-17T12:18:24.902 INFO:teuthology.orchestra.run.vm02.stdout:(33/151): ceph-mgr-modules-core-20.2.0-21.gc03b 2.5 MB/s | 289 kB 00:00 2026-04-17T12:18:24.951 INFO:teuthology.orchestra.run.vm02.stdout:(34/151): ceph-mgr-rook-20.2.0-21.gc03ba9ecf58. 1.0 MB/s | 50 kB 00:00 2026-04-17T12:18:24.979 INFO:teuthology.orchestra.run.vm02.stdout:(35/151): ceph-mgr-diskprediction-local-20.2.0- 17 MB/s | 7.4 MB 00:00 2026-04-17T12:18:25.058 INFO:teuthology.orchestra.run.vm02.stdout:(36/151): ceph-prometheus-alerts-20.2.0-21.gc03 160 kB/s | 17 kB 00:00 2026-04-17T12:18:25.061 INFO:teuthology.orchestra.run.vm02.stdout:(37/151): ceph-volume-20.2.0-21.gc03ba9ecf58.el 3.6 MB/s | 297 kB 00:00 2026-04-17T12:18:25.082 INFO:teuthology.orchestra.run.vm02.stdout:(38/151): abseil-cpp-20211102.0-4.el9.x86_64.rp 26 MB/s | 551 kB 00:00 2026-04-17T12:18:25.138 INFO:teuthology.orchestra.run.vm02.stdout:(39/151): gperftools-libs-2.9.1-3.el9.x86_64.rp 5.4 MB/s | 308 kB 00:00 2026-04-17T12:18:25.145 INFO:teuthology.orchestra.run.vm02.stdout:(40/151): grpc-data-1.46.7-10.el9.noarch.rpm 2.8 MB/s | 19 kB 00:00 2026-04-17T12:18:25.267 INFO:teuthology.orchestra.run.vm02.stdout:(41/151): libarrow-9.0.0-15.el9.x86_64.rpm 36 MB/s | 4.4 MB 00:00 2026-04-17T12:18:25.272 INFO:teuthology.orchestra.run.vm02.stdout:(42/151): libarrow-doc-9.0.0-15.el9.noarch.rpm 5.4 MB/s | 25 kB 00:00 2026-04-17T12:18:25.276 INFO:teuthology.orchestra.run.vm02.stdout:(43/151): liboath-2.6.12-1.el9.x86_64.rpm 12 MB/s | 49 kB 00:00 2026-04-17T12:18:25.282 INFO:teuthology.orchestra.run.vm02.stdout:(44/151): libunwind-1.6.2-1.el9.x86_64.rpm 12 MB/s | 67 kB 00:00 2026-04-17T12:18:25.287 INFO:teuthology.orchestra.run.vm02.stdout:(45/151): luarocks-3.9.2-5.el9.noarch.rpm 30 MB/s | 151 kB 00:00 2026-04-17T12:18:25.314 INFO:teuthology.orchestra.run.vm02.stdout:(46/151): parquet-libs-9.0.0-15.el9.x86_64.rpm 31 MB/s | 838 kB 00:00 2026-04-17T12:18:25.331 INFO:teuthology.orchestra.run.vm02.stdout:(47/151): python3-asyncssh-2.13.2-5.el9.noarch. 32 MB/s | 548 kB 00:00 2026-04-17T12:18:25.335 INFO:teuthology.orchestra.run.vm02.stdout:(48/151): python3-autocommand-2.2.2-8.el9.noarc 7.6 MB/s | 29 kB 00:00 2026-04-17T12:18:25.341 INFO:teuthology.orchestra.run.vm02.stdout:(49/151): python3-backports-tarfile-1.2.0-1.el9 11 MB/s | 60 kB 00:00 2026-04-17T12:18:25.345 INFO:teuthology.orchestra.run.vm02.stdout:(50/151): python3-bcrypt-3.2.2-1.el9.x86_64.rpm 12 MB/s | 43 kB 00:00 2026-04-17T12:18:25.347 INFO:teuthology.orchestra.run.vm02.stdout:(51/151): python3-cachetools-4.2.4-1.el9.noarch 12 MB/s | 32 kB 00:00 2026-04-17T12:18:25.351 INFO:teuthology.orchestra.run.vm02.stdout:(52/151): python3-certifi-2023.05.07-4.el9.noar 4.0 MB/s | 14 kB 00:00 2026-04-17T12:18:25.356 INFO:teuthology.orchestra.run.vm02.stdout:(53/151): python3-cheroot-10.0.1-5.el9.noarch.r 35 MB/s | 173 kB 00:00 2026-04-17T12:18:25.367 INFO:teuthology.orchestra.run.vm02.stdout:(54/151): python3-cherrypy-18.10.0-5.el9.noarch 28 MB/s | 290 kB 00:00 2026-04-17T12:18:25.373 INFO:teuthology.orchestra.run.vm02.stdout:(55/151): python3-google-auth-2.45.0-1.el9.noar 38 MB/s | 254 kB 00:00 2026-04-17T12:18:25.411 INFO:teuthology.orchestra.run.vm02.stdout:(56/151): cephadm-20.2.0-21.gc03ba9ecf58.el9.cl 2.8 MB/s | 1.0 MB 00:00 2026-04-17T12:18:25.433 INFO:teuthology.orchestra.run.vm02.stdout:(57/151): python3-grpcio-1.46.7-10.el9.x86_64.r 34 MB/s | 2.0 MB 00:00 2026-04-17T12:18:25.434 INFO:teuthology.orchestra.run.vm02.stdout:(58/151): python3-grpcio-tools-1.46.7-10.el9.x8 6.1 MB/s | 144 kB 00:00 2026-04-17T12:18:25.439 INFO:teuthology.orchestra.run.vm02.stdout:(59/151): python3-influxdb-5.3.1-1.el9.noarch.r 25 MB/s | 139 kB 00:00 2026-04-17T12:18:25.440 INFO:teuthology.orchestra.run.vm02.stdout:(60/151): python3-isodate-0.6.1-3.el9.noarch.rp 9.7 MB/s | 56 kB 00:00 2026-04-17T12:18:25.442 INFO:teuthology.orchestra.run.vm02.stdout:(61/151): python3-jaraco-8.2.1-3.el9.noarch.rpm 3.5 MB/s | 11 kB 00:00 2026-04-17T12:18:25.443 INFO:teuthology.orchestra.run.vm02.stdout:(62/151): python3-jaraco-classes-3.2.1-5.el9.no 6.6 MB/s | 18 kB 00:00 2026-04-17T12:18:25.446 INFO:teuthology.orchestra.run.vm02.stdout:(63/151): python3-jaraco-collections-3.0.0-8.el 5.7 MB/s | 23 kB 00:00 2026-04-17T12:18:25.447 INFO:teuthology.orchestra.run.vm02.stdout:(64/151): python3-jaraco-context-6.0.1-3.el9.no 5.1 MB/s | 20 kB 00:00 2026-04-17T12:18:25.451 INFO:teuthology.orchestra.run.vm02.stdout:(65/151): python3-jaraco-functools-3.5.0-2.el9. 4.6 MB/s | 19 kB 00:00 2026-04-17T12:18:25.452 INFO:teuthology.orchestra.run.vm02.stdout:(66/151): python3-jaraco-text-4.0.0-2.el9.noarc 5.2 MB/s | 26 kB 00:00 2026-04-17T12:18:25.461 INFO:teuthology.orchestra.run.vm02.stdout:(67/151): python3-more-itertools-8.12.0-2.el9.n 9.2 MB/s | 79 kB 00:00 2026-04-17T12:18:25.474 INFO:teuthology.orchestra.run.vm02.stdout:(68/151): python3-msgpack-1.0.3-2.el9.x86_64.rp 7.3 MB/s | 86 kB 00:00 2026-04-17T12:18:25.480 INFO:teuthology.orchestra.run.vm02.stdout:(69/151): python3-natsort-7.1.1-5.el9.noarch.rp 8.3 MB/s | 58 kB 00:00 2026-04-17T12:18:25.484 INFO:teuthology.orchestra.run.vm02.stdout:(70/151): python3-kubernetes-26.1.0-3.el9.noarc 31 MB/s | 1.0 MB 00:00 2026-04-17T12:18:25.486 INFO:teuthology.orchestra.run.vm02.stdout:(71/151): python3-portend-3.1.0-2.el9.noarch.rp 2.6 MB/s | 16 kB 00:00 2026-04-17T12:18:25.489 INFO:teuthology.orchestra.run.vm02.stdout:(72/151): python3-pyOpenSSL-21.0.0-1.el9.noarch 16 MB/s | 90 kB 00:00 2026-04-17T12:18:25.492 INFO:teuthology.orchestra.run.vm02.stdout:(73/151): python3-repoze-lru-0.7-16.el9.noarch. 5.7 MB/s | 31 kB 00:00 2026-04-17T12:18:25.496 INFO:teuthology.orchestra.run.vm02.stdout:(74/151): python3-routes-2.5.1-5.el9.noarch.rpm 30 MB/s | 188 kB 00:00 2026-04-17T12:18:25.498 INFO:teuthology.orchestra.run.vm02.stdout:(75/151): python3-rsa-4.9-2.el9.noarch.rpm 10 MB/s | 59 kB 00:00 2026-04-17T12:18:25.503 INFO:teuthology.orchestra.run.vm02.stdout:(76/151): python3-saml-1.16.0-1.el9.noarch.rpm 18 MB/s | 125 kB 00:00 2026-04-17T12:18:25.503 INFO:teuthology.orchestra.run.vm02.stdout:(77/151): python3-tempora-5.0.0-2.el9.noarch.rp 6.3 MB/s | 36 kB 00:00 2026-04-17T12:18:25.508 INFO:teuthology.orchestra.run.vm02.stdout:(78/151): python3-typing-extensions-4.15.0-1.el 16 MB/s | 86 kB 00:00 2026-04-17T12:18:25.510 INFO:teuthology.orchestra.run.vm02.stdout:(79/151): python3-websocket-client-1.2.3-2.el9. 13 MB/s | 90 kB 00:00 2026-04-17T12:18:25.513 INFO:teuthology.orchestra.run.vm02.stdout:(80/151): python3-xmlsec-1.3.13-1.el9.x86_64.rp 10 MB/s | 48 kB 00:00 2026-04-17T12:18:25.514 INFO:teuthology.orchestra.run.vm02.stdout:(81/151): python3-xmltodict-0.12.0-15.el9.noarc 5.3 MB/s | 22 kB 00:00 2026-04-17T12:18:25.516 INFO:teuthology.orchestra.run.vm02.stdout:(82/151): python3-zc-lockfile-2.0-10.el9.noarch 6.6 MB/s | 20 kB 00:00 2026-04-17T12:18:25.521 INFO:teuthology.orchestra.run.vm02.stdout:(83/151): re2-20211101-20.el9.x86_64.rpm 31 MB/s | 191 kB 00:00 2026-04-17T12:18:25.524 INFO:teuthology.orchestra.run.vm02.stdout:(84/151): s3cmd-2.4.0-1.el9.noarch.rpm 24 MB/s | 206 kB 00:00 2026-04-17T12:18:25.565 INFO:teuthology.orchestra.run.vm02.stdout:(85/151): thrift-0.15.0-4.el9.x86_64.rpm 36 MB/s | 1.6 MB 00:00 2026-04-17T12:18:25.661 INFO:teuthology.orchestra.run.vm02.stdout:(86/151): bzip2-1.0.8-10.el9_5.x86_64.rpm 377 kB/s | 51 kB 00:00 2026-04-17T12:18:25.723 INFO:teuthology.orchestra.run.vm02.stdout:(87/151): c-ares-1.19.1-2.el9_4.x86_64.rpm 697 kB/s | 110 kB 00:00 2026-04-17T12:18:25.740 INFO:teuthology.orchestra.run.vm02.stdout:(88/151): cryptsetup-2.7.2-4.el9.x86_64.rpm 3.8 MB/s | 310 kB 00:00 2026-04-17T12:18:25.751 INFO:teuthology.orchestra.run.vm02.stdout:(89/151): fuse-2.9.9-17.el9.x86_64.rpm 2.7 MB/s | 78 kB 00:00 2026-04-17T12:18:25.768 INFO:teuthology.orchestra.run.vm02.stdout:(90/151): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-17T12:18:25.779 INFO:teuthology.orchestra.run.vm02.stdout:(91/151): libconfig-1.7.2-9.el9.x86_64.rpm 2.5 MB/s | 71 kB 00:00 2026-04-17T12:18:25.805 INFO:teuthology.orchestra.run.vm06.stdout:(30/151): ceph-mgr-dashboard-20.2.0-21.gc03ba9e 13 MB/s | 15 MB 00:01 2026-04-17T12:18:25.826 INFO:teuthology.orchestra.run.vm02.stdout:(92/151): libgfortran-11.5.0-11.el9.x86_64.rpm 13 MB/s | 794 kB 00:00 2026-04-17T12:18:25.833 INFO:teuthology.orchestra.run.vm02.stdout:(93/151): libquadmath-11.5.0-11.el9.x86_64.rpm 3.4 MB/s | 184 kB 00:00 2026-04-17T12:18:25.854 INFO:teuthology.orchestra.run.vm02.stdout:(94/151): mailcap-2.1.49-5.el9.0.2.noarch.rpm 1.2 MB/s | 32 kB 00:00 2026-04-17T12:18:25.882 INFO:teuthology.orchestra.run.vm02.stdout:(95/151): pciutils-3.7.0-7.el9.x86_64.rpm 3.2 MB/s | 92 kB 00:00 2026-04-17T12:18:25.902 INFO:teuthology.orchestra.run.vm02.stdout:(96/151): nvme-cli-2.13-1.el9.x86_64.rpm 14 MB/s | 1.0 MB 00:00 2026-04-17T12:18:25.919 INFO:teuthology.orchestra.run.vm02.stdout:(97/151): python3-cffi-1.14.5-5.el9.x86_64.rpm 6.5 MB/s | 241 kB 00:00 2026-04-17T12:18:25.919 INFO:teuthology.orchestra.run.vm08.stdout:(10/151): ceph-radosgw-20.2.0-21.gc03ba9ecf58.e 6.1 MB/s | 24 MB 00:03 2026-04-17T12:18:25.945 INFO:teuthology.orchestra.run.vm02.stdout:(98/151): python3-cryptography-36.0.1-5.el9_6.x 27 MB/s | 1.2 MB 00:00 2026-04-17T12:18:25.947 INFO:teuthology.orchestra.run.vm02.stdout:(99/151): python3-ply-3.11-14.el9.0.1.noarch.rp 3.6 MB/s | 103 kB 00:00 2026-04-17T12:18:25.959 INFO:teuthology.orchestra.run.vm08.stdout:(11/151): libcephfs-daemon-20.2.0-21.gc03ba9ecf 917 kB/s | 35 kB 00:00 2026-04-17T12:18:25.974 INFO:teuthology.orchestra.run.vm02.stdout:(100/151): python3-pycparser-2.20-6.el9.noarch. 4.2 MB/s | 124 kB 00:00 2026-04-17T12:18:25.977 INFO:teuthology.orchestra.run.vm02.stdout:(101/151): python3-pyparsing-2.4.7-9.el9.0.1.no 4.9 MB/s | 150 kB 00:00 2026-04-17T12:18:26.003 INFO:teuthology.orchestra.run.vm02.stdout:(102/151): python3-requests-2.25.1-10.el9_6.noa 3.9 MB/s | 115 kB 00:00 2026-04-17T12:18:26.004 INFO:teuthology.orchestra.run.vm08.stdout:(12/151): libcephfs-devel-20.2.0-21.gc03ba9ecf5 751 kB/s | 34 kB 00:00 2026-04-17T12:18:26.007 INFO:teuthology.orchestra.run.vm02.stdout:(103/151): python3-urllib3-1.26.5-6.el9_7.1.noa 6.2 MB/s | 191 kB 00:00 2026-04-17T12:18:26.039 INFO:teuthology.orchestra.run.vm08.stdout:(13/151): libcephfs-proxy2-20.2.0-21.gc03ba9ecf 700 kB/s | 24 kB 00:00 2026-04-17T12:18:26.040 INFO:teuthology.orchestra.run.vm02.stdout:(104/151): smartmontools-7.2-9.el9.x86_64.rpm 15 MB/s | 551 kB 00:00 2026-04-17T12:18:26.041 INFO:teuthology.orchestra.run.vm02.stdout:(105/151): unzip-6.0-59.el9.x86_64.rpm 5.2 MB/s | 180 kB 00:00 2026-04-17T12:18:26.073 INFO:teuthology.orchestra.run.vm02.stdout:(106/151): zip-3.0-35.el9.x86_64.rpm 7.9 MB/s | 263 kB 00:00 2026-04-17T12:18:26.073 INFO:teuthology.orchestra.run.vm02.stdout:(107/151): boost-program-options-1.75.0-13.el9_ 3.2 MB/s | 104 kB 00:00 2026-04-17T12:18:26.103 INFO:teuthology.orchestra.run.vm02.stdout:(108/151): flexiblas-3.0.4-8.el9.0.1.x86_64.rpm 1.0 MB/s | 30 kB 00:00 2026-04-17T12:18:26.133 INFO:teuthology.orchestra.run.vm02.stdout:(109/151): flexiblas-openblas-openmp-3.0.4-8.el 512 kB/s | 15 kB 00:00 2026-04-17T12:18:26.168 INFO:teuthology.orchestra.run.vm02.stdout:(110/151): flexiblas-netlib-3.0.4-8.el9.0.1.x86 32 MB/s | 3.0 MB 00:00 2026-04-17T12:18:26.170 INFO:teuthology.orchestra.run.vm02.stdout:(111/151): libnbd-1.20.3-4.el9.x86_64.rpm 4.5 MB/s | 171 kB 00:00 2026-04-17T12:18:26.195 INFO:teuthology.orchestra.run.vm08.stdout:(14/151): libcephfs2-20.2.0-21.gc03ba9ecf58.el9 5.5 MB/s | 868 kB 00:00 2026-04-17T12:18:26.198 INFO:teuthology.orchestra.run.vm02.stdout:(112/151): libpmemobj-1.12.1-1.el9.x86_64.rpm 5.4 MB/s | 159 kB 00:00 2026-04-17T12:18:26.199 INFO:teuthology.orchestra.run.vm02.stdout:(113/151): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.5 MB/s | 44 kB 00:00 2026-04-17T12:18:26.235 INFO:teuthology.orchestra.run.vm02.stdout:(114/151): libstoragemgmt-1.10.1-1.el9.x86_64.r 6.7 MB/s | 243 kB 00:00 2026-04-17T12:18:26.240 INFO:teuthology.orchestra.run.vm02.stdout:(115/151): librdkafka-1.6.1-102.el9.x86_64.rpm 16 MB/s | 662 kB 00:00 2026-04-17T12:18:26.242 INFO:teuthology.orchestra.run.vm08.stdout:(15/151): libcephsqlite-20.2.0-21.gc03ba9ecf58. 3.4 MB/s | 164 kB 00:00 2026-04-17T12:18:26.264 INFO:teuthology.orchestra.run.vm02.stdout:(116/151): libxslt-1.1.34-14.el9_7.1.x86_64.rpm 8.2 MB/s | 240 kB 00:00 2026-04-17T12:18:26.271 INFO:teuthology.orchestra.run.vm02.stdout:(117/151): lttng-ust-2.12.0-6.el9.x86_64.rpm 8.8 MB/s | 282 kB 00:00 2026-04-17T12:18:26.293 INFO:teuthology.orchestra.run.vm02.stdout:(118/151): lua-5.4.4-4.el9.x86_64.rpm 6.4 MB/s | 187 kB 00:00 2026-04-17T12:18:26.299 INFO:teuthology.orchestra.run.vm02.stdout:(119/151): openblas-0.3.29-1.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-17T12:18:26.300 INFO:teuthology.orchestra.run.vm08.stdout:(16/151): librados-devel-20.2.0-21.gc03ba9ecf58 2.1 MB/s | 126 kB 00:00 2026-04-17T12:18:26.330 INFO:teuthology.orchestra.run.vm02.stdout:(120/151): perl-Benchmark-1.23-481.1.el9_6.noar 838 kB/s | 25 kB 00:00 2026-04-17T12:18:26.364 INFO:teuthology.orchestra.run.vm02.stdout:(121/151): perl-Test-Harness-3.42-461.el9.noarc 7.8 MB/s | 267 kB 00:00 2026-04-17T12:18:26.370 INFO:teuthology.orchestra.run.vm08.stdout:(17/151): libradosstriper1-20.2.0-21.gc03ba9ecf 3.6 MB/s | 250 kB 00:00 2026-04-17T12:18:26.407 INFO:teuthology.orchestra.run.vm02.stdout:(122/151): protobuf-3.14.0-17.el9_7.x86_64.rpm 23 MB/s | 1.0 MB 00:00 2026-04-17T12:18:26.482 INFO:teuthology.orchestra.run.vm02.stdout:(123/151): openblas-openmp-0.3.29-1.el9.x86_64. 28 MB/s | 5.3 MB 00:00 2026-04-17T12:18:26.511 INFO:teuthology.orchestra.run.vm02.stdout:(124/151): python3-devel-3.9.23-2.el9.x86_64.rp 7.0 MB/s | 205 kB 00:00 2026-04-17T12:18:26.543 INFO:teuthology.orchestra.run.vm02.stdout:(125/151): python3-jinja2-2.11.3-8.el9_5.noarch 7.2 MB/s | 228 kB 00:00 2026-04-17T12:18:26.557 INFO:teuthology.orchestra.run.vm02.stdout:(126/151): python3-babel-2.9.1-2.el9.noarch.rpm 39 MB/s | 5.8 MB 00:00 2026-04-17T12:18:26.570 INFO:teuthology.orchestra.run.vm02.stdout:(127/151): python3-jmespath-1.0.1-1.el9_7.noarc 1.6 MB/s | 43 kB 00:00 2026-04-17T12:18:26.585 INFO:teuthology.orchestra.run.vm02.stdout:(128/151): python3-libstoragemgmt-1.10.1-1.el9. 5.8 MB/s | 166 kB 00:00 2026-04-17T12:18:26.611 INFO:teuthology.orchestra.run.vm02.stdout:(129/151): python3-markupsafe-1.1.1-12.el9.x86_ 1.2 MB/s | 32 kB 00:00 2026-04-17T12:18:26.630 INFO:teuthology.orchestra.run.vm02.stdout:(130/151): python3-lxml-4.6.5-3.el9.x86_64.rpm 20 MB/s | 1.2 MB 00:00 2026-04-17T12:18:26.664 INFO:teuthology.orchestra.run.vm02.stdout:(131/151): python3-numpy-f2py-1.23.5-2.el9_7.x8 11 MB/s | 368 kB 00:00 2026-04-17T12:18:26.700 INFO:teuthology.orchestra.run.vm02.stdout:(132/151): python3-packaging-20.9-5.el9.noarch. 1.9 MB/s | 69 kB 00:00 2026-04-17T12:18:26.733 INFO:teuthology.orchestra.run.vm02.stdout:(133/151): python3-protobuf-3.14.0-17.el9_7.noa 7.1 MB/s | 237 kB 00:00 2026-04-17T12:18:26.753 INFO:teuthology.orchestra.run.vm02.stdout:(134/151): python3-numpy-1.23.5-2.el9_7.x86_64. 41 MB/s | 5.8 MB 00:00 2026-04-17T12:18:26.761 INFO:teuthology.orchestra.run.vm02.stdout:(135/151): python3-pyasn1-0.4.8-7.el9_7.noarch. 4.6 MB/s | 132 kB 00:00 2026-04-17T12:18:26.785 INFO:teuthology.orchestra.run.vm02.stdout:(136/151): python3-pyasn1-modules-0.4.8-7.el9_7 6.6 MB/s | 210 kB 00:00 2026-04-17T12:18:26.785 INFO:teuthology.orchestra.run.vm06.stdout:(31/151): ceph-mgr-diskprediction-local-20.2.0- 7.5 MB/s | 7.4 MB 00:00 2026-04-17T12:18:26.789 INFO:teuthology.orchestra.run.vm02.stdout:(137/151): python3-requests-oauthlib-1.3.0-12.e 1.6 MB/s | 43 kB 00:00 2026-04-17T12:18:26.818 INFO:teuthology.orchestra.run.vm06.stdout:(32/151): ceph-mgr-k8sevents-20.2.0-21.gc03ba9e 691 kB/s | 22 kB 00:00 2026-04-17T12:18:26.818 INFO:teuthology.orchestra.run.vm02.stdout:(138/151): python3-toml-0.10.2-6.el9.0.1.noarch 1.5 MB/s | 44 kB 00:00 2026-04-17T12:18:26.849 INFO:teuthology.orchestra.run.vm02.stdout:(139/151): qatlib-24.09.0-1.el9.x86_64.rpm 7.1 MB/s | 221 kB 00:00 2026-04-17T12:18:26.881 INFO:teuthology.orchestra.run.vm02.stdout:(140/151): qatlib-service-24.09.0-1.el9.x86_64. 1.1 MB/s | 36 kB 00:00 2026-04-17T12:18:26.883 INFO:teuthology.orchestra.run.vm06.stdout:(33/151): ceph-mgr-modules-core-20.2.0-21.gc03b 4.3 MB/s | 289 kB 00:00 2026-04-17T12:18:26.910 INFO:teuthology.orchestra.run.vm02.stdout:(141/151): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.2 MB/s | 65 kB 00:00 2026-04-17T12:18:26.915 INFO:teuthology.orchestra.run.vm06.stdout:(34/151): ceph-mgr-rook-20.2.0-21.gc03ba9ecf58. 1.6 MB/s | 50 kB 00:00 2026-04-17T12:18:26.944 INFO:teuthology.orchestra.run.vm02.stdout:(142/151): socat-1.7.4.1-8.el9.x86_64.rpm 8.8 MB/s | 299 kB 00:00 2026-04-17T12:18:26.946 INFO:teuthology.orchestra.run.vm06.stdout:(35/151): ceph-prometheus-alerts-20.2.0-21.gc03 551 kB/s | 17 kB 00:00 2026-04-17T12:18:26.975 INFO:teuthology.orchestra.run.vm02.stdout:(143/151): xmlsec1-1.2.29-13.el9.x86_64.rpm 5.9 MB/s | 188 kB 00:00 2026-04-17T12:18:27.017 INFO:teuthology.orchestra.run.vm02.stdout:(144/151): xmlsec1-openssl-1.2.29-13.el9.x86_64 2.1 MB/s | 89 kB 00:00 2026-04-17T12:18:27.017 INFO:teuthology.orchestra.run.vm08.stdout:(18/151): ceph-osd-20.2.0-21.gc03ba9ecf58.el9.c 3.3 MB/s | 17 MB 00:05 2026-04-17T12:18:27.030 INFO:teuthology.orchestra.run.vm06.stdout:(36/151): ceph-volume-20.2.0-21.gc03ba9ecf58.el 3.5 MB/s | 297 kB 00:00 2026-04-17T12:18:27.048 INFO:teuthology.orchestra.run.vm02.stdout:(145/151): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.0 MB/s | 63 kB 00:00 2026-04-17T12:18:27.060 INFO:teuthology.orchestra.run.vm08.stdout:(19/151): python3-ceph-argparse-20.2.0-21.gc03b 1.0 MB/s | 45 kB 00:00 2026-04-17T12:18:27.077 INFO:teuthology.orchestra.run.vm02.stdout:(146/151): lua-devel-5.4.4-4.el9.x86_64.rpm 747 kB/s | 21 kB 00:00 2026-04-17T12:18:27.169 INFO:teuthology.orchestra.run.vm02.stdout:(147/151): protobuf-compiler-3.14.0-17.el9_7.x8 9.2 MB/s | 862 kB 00:00 2026-04-17T12:18:27.215 INFO:teuthology.orchestra.run.vm06.stdout:(37/151): cephadm-20.2.0-21.gc03ba9ecf58.el9.cl 5.3 MB/s | 1.0 MB 00:00 2026-04-17T12:18:27.215 INFO:teuthology.orchestra.run.vm02.stdout:(148/151): python3-scipy-1.9.3-2.el9.x86_64.rpm 44 MB/s | 19 MB 00:00 2026-04-17T12:18:27.244 INFO:teuthology.orchestra.run.vm08.stdout:(20/151): python3-ceph-common-20.2.0-21.gc03ba9 886 kB/s | 163 kB 00:00 2026-04-17T12:18:27.298 INFO:teuthology.orchestra.run.vm06.stdout:(38/151): abseil-cpp-20211102.0-4.el9.x86_64.rp 6.6 MB/s | 551 kB 00:00 2026-04-17T12:18:27.304 INFO:teuthology.orchestra.run.vm06.stdout:(39/151): gperftools-libs-2.9.1-3.el9.x86_64.rp 48 MB/s | 308 kB 00:00 2026-04-17T12:18:27.308 INFO:teuthology.orchestra.run.vm06.stdout:(40/151): grpc-data-1.46.7-10.el9.noarch.rpm 6.4 MB/s | 19 kB 00:00 2026-04-17T12:18:27.355 INFO:teuthology.orchestra.run.vm08.stdout:(21/151): python3-cephfs-20.2.0-21.gc03ba9ecf58 1.4 MB/s | 163 kB 00:00 2026-04-17T12:18:27.386 INFO:teuthology.orchestra.run.vm06.stdout:(41/151): libarrow-9.0.0-15.el9.x86_64.rpm 57 MB/s | 4.4 MB 00:00 2026-04-17T12:18:27.390 INFO:teuthology.orchestra.run.vm06.stdout:(42/151): libarrow-doc-9.0.0-15.el9.noarch.rpm 6.8 MB/s | 25 kB 00:00 2026-04-17T12:18:27.396 INFO:teuthology.orchestra.run.vm06.stdout:(43/151): liboath-2.6.12-1.el9.x86_64.rpm 10 MB/s | 49 kB 00:00 2026-04-17T12:18:27.402 INFO:teuthology.orchestra.run.vm06.stdout:(44/151): libunwind-1.6.2-1.el9.x86_64.rpm 12 MB/s | 67 kB 00:00 2026-04-17T12:18:27.407 INFO:teuthology.orchestra.run.vm06.stdout:(45/151): luarocks-3.9.2-5.el9.noarch.rpm 29 MB/s | 151 kB 00:00 2026-04-17T12:18:27.424 INFO:teuthology.orchestra.run.vm08.stdout:(22/151): librgw2-20.2.0-21.gc03ba9ecf58.el9.cl 6.0 MB/s | 6.4 MB 00:01 2026-04-17T12:18:27.434 INFO:teuthology.orchestra.run.vm06.stdout:(46/151): parquet-libs-9.0.0-15.el9.x86_64.rpm 31 MB/s | 838 kB 00:00 2026-04-17T12:18:27.446 INFO:teuthology.orchestra.run.vm06.stdout:(47/151): python3-asyncssh-2.13.2-5.el9.noarch. 47 MB/s | 548 kB 00:00 2026-04-17T12:18:27.450 INFO:teuthology.orchestra.run.vm06.stdout:(48/151): python3-autocommand-2.2.2-8.el9.noarc 7.9 MB/s | 29 kB 00:00 2026-04-17T12:18:27.456 INFO:teuthology.orchestra.run.vm06.stdout:(49/151): python3-backports-tarfile-1.2.0-1.el9 11 MB/s | 60 kB 00:00 2026-04-17T12:18:27.460 INFO:teuthology.orchestra.run.vm06.stdout:(50/151): python3-bcrypt-3.2.2-1.el9.x86_64.rpm 13 MB/s | 43 kB 00:00 2026-04-17T12:18:27.464 INFO:teuthology.orchestra.run.vm06.stdout:(51/151): python3-cachetools-4.2.4-1.el9.noarch 8.7 MB/s | 32 kB 00:00 2026-04-17T12:18:27.468 INFO:teuthology.orchestra.run.vm06.stdout:(52/151): python3-certifi-2023.05.07-4.el9.noar 3.6 MB/s | 14 kB 00:00 2026-04-17T12:18:27.475 INFO:teuthology.orchestra.run.vm06.stdout:(53/151): python3-cheroot-10.0.1-5.el9.noarch.r 29 MB/s | 173 kB 00:00 2026-04-17T12:18:27.483 INFO:teuthology.orchestra.run.vm06.stdout:(54/151): python3-cherrypy-18.10.0-5.el9.noarch 37 MB/s | 290 kB 00:00 2026-04-17T12:18:27.491 INFO:teuthology.orchestra.run.vm06.stdout:(55/151): python3-google-auth-2.45.0-1.el9.noar 32 MB/s | 254 kB 00:00 2026-04-17T12:18:27.509 INFO:teuthology.orchestra.run.vm08.stdout:(23/151): python3-rbd-20.2.0-21.gc03ba9ecf58.el 3.5 MB/s | 305 kB 00:00 2026-04-17T12:18:27.526 INFO:teuthology.orchestra.run.vm08.stdout:(24/151): python3-rados-20.2.0-21.gc03ba9ecf58. 1.8 MB/s | 317 kB 00:00 2026-04-17T12:18:27.531 INFO:teuthology.orchestra.run.vm06.stdout:(56/151): python3-grpcio-1.46.7-10.el9.x86_64.r 52 MB/s | 2.0 MB 00:00 2026-04-17T12:18:27.536 INFO:teuthology.orchestra.run.vm06.stdout:(57/151): python3-grpcio-tools-1.46.7-10.el9.x8 30 MB/s | 144 kB 00:00 2026-04-17T12:18:27.541 INFO:teuthology.orchestra.run.vm06.stdout:(58/151): python3-influxdb-5.3.1-1.el9.noarch.r 30 MB/s | 139 kB 00:00 2026-04-17T12:18:27.543 INFO:teuthology.orchestra.run.vm08.stdout:(25/151): python3-rgw-20.2.0-21.gc03ba9ecf58.el 2.9 MB/s | 99 kB 00:00 2026-04-17T12:18:27.545 INFO:teuthology.orchestra.run.vm06.stdout:(59/151): python3-isodate-0.6.1-3.el9.noarch.rp 15 MB/s | 56 kB 00:00 2026-04-17T12:18:27.549 INFO:teuthology.orchestra.run.vm06.stdout:(60/151): python3-jaraco-8.2.1-3.el9.noarch.rpm 3.8 MB/s | 11 kB 00:00 2026-04-17T12:18:27.554 INFO:teuthology.orchestra.run.vm06.stdout:(61/151): python3-jaraco-classes-3.2.1-5.el9.no 3.7 MB/s | 18 kB 00:00 2026-04-17T12:18:27.554 INFO:teuthology.orchestra.run.vm02.stdout:(149/151): librbd1-20.2.0-21.gc03ba9ecf58.el9.c 8.4 MB/s | 2.8 MB 00:00 2026-04-17T12:18:27.557 INFO:teuthology.orchestra.run.vm06.stdout:(62/151): python3-jaraco-collections-3.0.0-8.el 8.2 MB/s | 23 kB 00:00 2026-04-17T12:18:27.560 INFO:teuthology.orchestra.run.vm06.stdout:(63/151): python3-jaraco-context-6.0.1-3.el9.no 6.2 MB/s | 20 kB 00:00 2026-04-17T12:18:27.564 INFO:teuthology.orchestra.run.vm06.stdout:(64/151): python3-jaraco-functools-3.5.0-2.el9. 5.7 MB/s | 19 kB 00:00 2026-04-17T12:18:27.568 INFO:teuthology.orchestra.run.vm06.stdout:(65/151): python3-jaraco-text-4.0.0-2.el9.noarc 7.7 MB/s | 26 kB 00:00 2026-04-17T12:18:27.584 INFO:teuthology.orchestra.run.vm08.stdout:(26/151): rbd-fuse-20.2.0-21.gc03ba9ecf58.el9.c 1.5 MB/s | 91 kB 00:00 2026-04-17T12:18:27.585 INFO:teuthology.orchestra.run.vm06.stdout:(66/151): python3-kubernetes-26.1.0-3.el9.noarc 61 MB/s | 1.0 MB 00:00 2026-04-17T12:18:27.590 INFO:teuthology.orchestra.run.vm06.stdout:(67/151): python3-more-itertools-8.12.0-2.el9.n 17 MB/s | 79 kB 00:00 2026-04-17T12:18:27.594 INFO:teuthology.orchestra.run.vm06.stdout:(68/151): python3-msgpack-1.0.3-2.el9.x86_64.rp 21 MB/s | 86 kB 00:00 2026-04-17T12:18:27.600 INFO:teuthology.orchestra.run.vm06.stdout:(69/151): python3-natsort-7.1.1-5.el9.noarch.rp 9.5 MB/s | 58 kB 00:00 2026-04-17T12:18:27.605 INFO:teuthology.orchestra.run.vm06.stdout:(70/151): python3-portend-3.1.0-2.el9.noarch.rp 3.8 MB/s | 16 kB 00:00 2026-04-17T12:18:27.610 INFO:teuthology.orchestra.run.vm06.stdout:(71/151): python3-pyOpenSSL-21.0.0-1.el9.noarch 18 MB/s | 90 kB 00:00 2026-04-17T12:18:27.614 INFO:teuthology.orchestra.run.vm06.stdout:(72/151): python3-repoze-lru-0.7-16.el9.noarch. 8.6 MB/s | 31 kB 00:00 2026-04-17T12:18:27.622 INFO:teuthology.orchestra.run.vm06.stdout:(73/151): python3-routes-2.5.1-5.el9.noarch.rpm 24 MB/s | 188 kB 00:00 2026-04-17T12:18:27.627 INFO:teuthology.orchestra.run.vm06.stdout:(74/151): python3-rsa-4.9-2.el9.noarch.rpm 12 MB/s | 59 kB 00:00 2026-04-17T12:18:27.647 INFO:teuthology.orchestra.run.vm06.stdout:(75/151): python3-saml-1.16.0-1.el9.noarch.rpm 6.4 MB/s | 125 kB 00:00 2026-04-17T12:18:27.651 INFO:teuthology.orchestra.run.vm06.stdout:(76/151): python3-tempora-5.0.0-2.el9.noarch.rp 9.0 MB/s | 36 kB 00:00 2026-04-17T12:18:27.656 INFO:teuthology.orchestra.run.vm06.stdout:(77/151): python3-typing-extensions-4.15.0-1.el 18 MB/s | 86 kB 00:00 2026-04-17T12:18:27.660 INFO:teuthology.orchestra.run.vm06.stdout:(78/151): python3-websocket-client-1.2.3-2.el9. 21 MB/s | 90 kB 00:00 2026-04-17T12:18:27.663 INFO:teuthology.orchestra.run.vm06.stdout:(79/151): python3-xmlsec-1.3.13-1.el9.x86_64.rp 14 MB/s | 48 kB 00:00 2026-04-17T12:18:27.667 INFO:teuthology.orchestra.run.vm06.stdout:(80/151): python3-xmltodict-0.12.0-15.el9.noarc 6.5 MB/s | 22 kB 00:00 2026-04-17T12:18:27.669 INFO:teuthology.orchestra.run.vm06.stdout:(81/151): python3-zc-lockfile-2.0-10.el9.noarch 9.0 MB/s | 20 kB 00:00 2026-04-17T12:18:27.675 INFO:teuthology.orchestra.run.vm06.stdout:(82/151): re2-20211101-20.el9.x86_64.rpm 33 MB/s | 191 kB 00:00 2026-04-17T12:18:27.682 INFO:teuthology.orchestra.run.vm06.stdout:(83/151): s3cmd-2.4.0-1.el9.noarch.rpm 31 MB/s | 206 kB 00:00 2026-04-17T12:18:27.706 INFO:teuthology.orchestra.run.vm08.stdout:(27/151): rbd-nbd-20.2.0-21.gc03ba9ecf58.el9.cl 1.4 MB/s | 180 kB 00:00 2026-04-17T12:18:27.710 INFO:teuthology.orchestra.run.vm06.stdout:(84/151): thrift-0.15.0-4.el9.x86_64.rpm 56 MB/s | 1.6 MB 00:00 2026-04-17T12:18:27.737 INFO:teuthology.orchestra.run.vm08.stdout:(28/151): ceph-grafana-dashboards-20.2.0-21.gc0 1.4 MB/s | 43 kB 00:00 2026-04-17T12:18:27.842 INFO:teuthology.orchestra.run.vm08.stdout:(29/151): ceph-mgr-cephadm-20.2.0-21.gc03ba9ecf 1.6 MB/s | 173 kB 00:00 2026-04-17T12:18:27.845 INFO:teuthology.orchestra.run.vm06.stdout:(85/151): bzip2-1.0.8-10.el9_5.x86_64.rpm 383 kB/s | 51 kB 00:00 2026-04-17T12:18:27.899 INFO:teuthology.orchestra.run.vm06.stdout:(86/151): c-ares-1.19.1-2.el9_4.x86_64.rpm 2.0 MB/s | 110 kB 00:00 2026-04-17T12:18:27.954 INFO:teuthology.orchestra.run.vm06.stdout:(87/151): cryptsetup-2.7.2-4.el9.x86_64.rpm 5.5 MB/s | 310 kB 00:00 2026-04-17T12:18:27.983 INFO:teuthology.orchestra.run.vm06.stdout:(88/151): fuse-2.9.9-17.el9.x86_64.rpm 2.7 MB/s | 78 kB 00:00 2026-04-17T12:18:28.012 INFO:teuthology.orchestra.run.vm06.stdout:(89/151): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.4 MB/s | 41 kB 00:00 2026-04-17T12:18:28.040 INFO:teuthology.orchestra.run.vm06.stdout:(90/151): libconfig-1.7.2-9.el9.x86_64.rpm 2.5 MB/s | 71 kB 00:00 2026-04-17T12:18:28.098 INFO:teuthology.orchestra.run.vm06.stdout:(91/151): libgfortran-11.5.0-11.el9.x86_64.rpm 14 MB/s | 794 kB 00:00 2026-04-17T12:18:28.103 INFO:teuthology.orchestra.run.vm08.stdout:(30/151): rbd-mirror-20.2.0-21.gc03ba9ecf58.el9 5.2 MB/s | 2.9 MB 00:00 2026-04-17T12:18:28.127 INFO:teuthology.orchestra.run.vm06.stdout:(92/151): libquadmath-11.5.0-11.el9.x86_64.rpm 6.4 MB/s | 184 kB 00:00 2026-04-17T12:18:28.155 INFO:teuthology.orchestra.run.vm06.stdout:(93/151): mailcap-2.1.49-5.el9.0.2.noarch.rpm 1.2 MB/s | 32 kB 00:00 2026-04-17T12:18:28.202 INFO:teuthology.orchestra.run.vm06.stdout:(94/151): nvme-cli-2.13-1.el9.x86_64.rpm 21 MB/s | 1.0 MB 00:00 2026-04-17T12:18:28.210 INFO:teuthology.orchestra.run.vm02.stdout:(150/151): librados2-20.2.0-21.gc03ba9ecf58.el9 3.4 MB/s | 3.5 MB 00:01 2026-04-17T12:18:28.230 INFO:teuthology.orchestra.run.vm06.stdout:(95/151): pciutils-3.7.0-7.el9.x86_64.rpm 3.3 MB/s | 92 kB 00:00 2026-04-17T12:18:28.259 INFO:teuthology.orchestra.run.vm06.stdout:(96/151): python3-cffi-1.14.5-5.el9.x86_64.rpm 8.2 MB/s | 241 kB 00:00 2026-04-17T12:18:28.309 INFO:teuthology.orchestra.run.vm06.stdout:(97/151): python3-cryptography-36.0.1-5.el9_6.x 23 MB/s | 1.2 MB 00:00 2026-04-17T12:18:28.337 INFO:teuthology.orchestra.run.vm06.stdout:(98/151): python3-ply-3.11-14.el9.0.1.noarch.rp 3.7 MB/s | 103 kB 00:00 2026-04-17T12:18:28.366 INFO:teuthology.orchestra.run.vm06.stdout:(99/151): python3-pycparser-2.20-6.el9.noarch.r 4.3 MB/s | 124 kB 00:00 2026-04-17T12:18:28.397 INFO:teuthology.orchestra.run.vm06.stdout:(100/151): python3-pyparsing-2.4.7-9.el9.0.1.no 4.7 MB/s | 150 kB 00:00 2026-04-17T12:18:28.426 INFO:teuthology.orchestra.run.vm06.stdout:(101/151): python3-requests-2.25.1-10.el9_6.noa 4.0 MB/s | 115 kB 00:00 2026-04-17T12:18:28.456 INFO:teuthology.orchestra.run.vm06.stdout:(102/151): python3-urllib3-1.26.5-6.el9_7.1.noa 6.4 MB/s | 191 kB 00:00 2026-04-17T12:18:28.503 INFO:teuthology.orchestra.run.vm06.stdout:(103/151): smartmontools-7.2-9.el9.x86_64.rpm 12 MB/s | 551 kB 00:00 2026-04-17T12:18:28.535 INFO:teuthology.orchestra.run.vm06.stdout:(104/151): unzip-6.0-59.el9.x86_64.rpm 5.5 MB/s | 180 kB 00:00 2026-04-17T12:18:28.569 INFO:teuthology.orchestra.run.vm06.stdout:(105/151): zip-3.0-35.el9.x86_64.rpm 7.7 MB/s | 263 kB 00:00 2026-04-17T12:18:28.596 INFO:teuthology.orchestra.run.vm06.stdout:(106/151): boost-program-options-1.75.0-13.el9_ 3.7 MB/s | 104 kB 00:00 2026-04-17T12:18:28.624 INFO:teuthology.orchestra.run.vm06.stdout:(107/151): flexiblas-3.0.4-8.el9.0.1.x86_64.rpm 1.1 MB/s | 30 kB 00:00 2026-04-17T12:18:28.711 INFO:teuthology.orchestra.run.vm06.stdout:(108/151): flexiblas-netlib-3.0.4-8.el9.0.1.x86 35 MB/s | 3.0 MB 00:00 2026-04-17T12:18:28.738 INFO:teuthology.orchestra.run.vm06.stdout:(109/151): flexiblas-openblas-openmp-3.0.4-8.el 579 kB/s | 15 kB 00:00 2026-04-17T12:18:28.767 INFO:teuthology.orchestra.run.vm06.stdout:(110/151): libnbd-1.20.3-4.el9.x86_64.rpm 5.9 MB/s | 171 kB 00:00 2026-04-17T12:18:28.797 INFO:teuthology.orchestra.run.vm06.stdout:(111/151): libpmemobj-1.12.1-1.el9.x86_64.rpm 5.2 MB/s | 159 kB 00:00 2026-04-17T12:18:28.826 INFO:teuthology.orchestra.run.vm06.stdout:(112/151): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.5 MB/s | 44 kB 00:00 2026-04-17T12:18:28.864 INFO:teuthology.orchestra.run.vm06.stdout:(113/151): librdkafka-1.6.1-102.el9.x86_64.rpm 17 MB/s | 662 kB 00:00 2026-04-17T12:18:28.896 INFO:teuthology.orchestra.run.vm06.stdout:(114/151): libstoragemgmt-1.10.1-1.el9.x86_64.r 7.6 MB/s | 243 kB 00:00 2026-04-17T12:18:28.927 INFO:teuthology.orchestra.run.vm06.stdout:(115/151): libxslt-1.1.34-14.el9_7.1.x86_64.rpm 7.7 MB/s | 240 kB 00:00 2026-04-17T12:18:28.957 INFO:teuthology.orchestra.run.vm06.stdout:(116/151): lttng-ust-2.12.0-6.el9.x86_64.rpm 9.4 MB/s | 282 kB 00:00 2026-04-17T12:18:28.987 INFO:teuthology.orchestra.run.vm06.stdout:(117/151): lua-5.4.4-4.el9.x86_64.rpm 6.0 MB/s | 187 kB 00:00 2026-04-17T12:18:29.015 INFO:teuthology.orchestra.run.vm06.stdout:(118/151): openblas-0.3.29-1.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-17T12:18:29.311 INFO:teuthology.orchestra.run.vm08.stdout:(31/151): ceph-mgr-diskprediction-local-20.2.0- 6.1 MB/s | 7.4 MB 00:01 2026-04-17T12:18:29.314 INFO:teuthology.orchestra.run.vm06.stdout:(119/151): ceph-common-20.2.0-21.gc03ba9ecf58.e 2.5 MB/s | 24 MB 00:09 2026-04-17T12:18:29.330 INFO:teuthology.orchestra.run.vm06.stdout:(120/151): openblas-openmp-0.3.29-1.el9.x86_64. 17 MB/s | 5.3 MB 00:00 2026-04-17T12:18:29.343 INFO:teuthology.orchestra.run.vm08.stdout:(32/151): ceph-mgr-k8sevents-20.2.0-21.gc03ba9e 693 kB/s | 22 kB 00:00 2026-04-17T12:18:29.358 INFO:teuthology.orchestra.run.vm06.stdout:(121/151): perl-Test-Harness-3.42-461.el9.noarc 9.1 MB/s | 267 kB 00:00 2026-04-17T12:18:29.404 INFO:teuthology.orchestra.run.vm06.stdout:(122/151): protobuf-3.14.0-17.el9_7.x86_64.rpm 22 MB/s | 1.0 MB 00:00 2026-04-17T12:18:29.415 INFO:teuthology.orchestra.run.vm08.stdout:(33/151): ceph-mgr-modules-core-20.2.0-21.gc03b 4.0 MB/s | 289 kB 00:00 2026-04-17T12:18:29.425 INFO:teuthology.orchestra.run.vm06.stdout:(123/151): perl-Benchmark-1.23-481.1.el9_6.noar 231 kB/s | 25 kB 00:00 2026-04-17T12:18:29.450 INFO:teuthology.orchestra.run.vm08.stdout:(34/151): ceph-mgr-rook-20.2.0-21.gc03ba9ecf58. 1.4 MB/s | 50 kB 00:00 2026-04-17T12:18:29.480 INFO:teuthology.orchestra.run.vm08.stdout:(35/151): ceph-prometheus-alerts-20.2.0-21.gc03 585 kB/s | 17 kB 00:00 2026-04-17T12:18:29.531 INFO:teuthology.orchestra.run.vm06.stdout:(124/151): python3-devel-3.9.23-2.el9.x86_64.rp 1.9 MB/s | 205 kB 00:00 2026-04-17T12:18:29.547 INFO:teuthology.orchestra.run.vm08.stdout:(36/151): ceph-volume-20.2.0-21.gc03ba9ecf58.el 4.3 MB/s | 297 kB 00:00 2026-04-17T12:18:29.547 INFO:teuthology.orchestra.run.vm06.stdout:(125/151): python3-babel-2.9.1-2.el9.noarch.rpm 41 MB/s | 5.8 MB 00:00 2026-04-17T12:18:29.562 INFO:teuthology.orchestra.run.vm06.stdout:(126/151): python3-jinja2-2.11.3-8.el9_5.noarch 7.3 MB/s | 228 kB 00:00 2026-04-17T12:18:29.576 INFO:teuthology.orchestra.run.vm06.stdout:(127/151): python3-jmespath-1.0.1-1.el9_7.noarc 1.5 MB/s | 43 kB 00:00 2026-04-17T12:18:29.593 INFO:teuthology.orchestra.run.vm06.stdout:(128/151): python3-libstoragemgmt-1.10.1-1.el9. 5.3 MB/s | 166 kB 00:00 2026-04-17T12:18:29.620 INFO:teuthology.orchestra.run.vm06.stdout:(129/151): python3-markupsafe-1.1.1-12.el9.x86_ 1.2 MB/s | 32 kB 00:00 2026-04-17T12:18:29.632 INFO:teuthology.orchestra.run.vm06.stdout:(130/151): python3-lxml-4.6.5-3.el9.x86_64.rpm 21 MB/s | 1.2 MB 00:00 2026-04-17T12:18:29.667 INFO:teuthology.orchestra.run.vm06.stdout:(131/151): python3-numpy-f2py-1.23.5-2.el9_7.x8 10 MB/s | 368 kB 00:00 2026-04-17T12:18:29.695 INFO:teuthology.orchestra.run.vm06.stdout:(132/151): python3-packaging-20.9-5.el9.noarch. 2.5 MB/s | 69 kB 00:00 2026-04-17T12:18:29.738 INFO:teuthology.orchestra.run.vm06.stdout:(133/151): python3-protobuf-3.14.0-17.el9_7.noa 5.4 MB/s | 237 kB 00:00 2026-04-17T12:18:29.769 INFO:teuthology.orchestra.run.vm06.stdout:(134/151): python3-pyasn1-0.4.8-7.el9_7.noarch. 4.3 MB/s | 132 kB 00:00 2026-04-17T12:18:29.786 INFO:teuthology.orchestra.run.vm08.stdout:(37/151): cephadm-20.2.0-21.gc03ba9ecf58.el9.cl 4.1 MB/s | 1.0 MB 00:00 2026-04-17T12:18:29.823 INFO:teuthology.orchestra.run.vm06.stdout:(135/151): python3-numpy-1.23.5-2.el9_7.x86_64. 29 MB/s | 5.8 MB 00:00 2026-04-17T12:18:29.824 INFO:teuthology.orchestra.run.vm08.stdout:(38/151): abseil-cpp-20211102.0-4.el9.x86_64.rp 14 MB/s | 551 kB 00:00 2026-04-17T12:18:29.850 INFO:teuthology.orchestra.run.vm06.stdout:(136/151): python3-pyasn1-modules-0.4.8-7.el9_7 2.5 MB/s | 210 kB 00:00 2026-04-17T12:18:29.852 INFO:teuthology.orchestra.run.vm06.stdout:(137/151): python3-requests-oauthlib-1.3.0-12.e 1.5 MB/s | 43 kB 00:00 2026-04-17T12:18:29.855 INFO:teuthology.orchestra.run.vm08.stdout:(39/151): gperftools-libs-2.9.1-3.el9.x86_64.rp 9.7 MB/s | 308 kB 00:00 2026-04-17T12:18:29.861 INFO:teuthology.orchestra.run.vm08.stdout:(40/151): grpc-data-1.46.7-10.el9.noarch.rpm 3.7 MB/s | 19 kB 00:00 2026-04-17T12:18:29.913 INFO:teuthology.orchestra.run.vm06.stdout:(138/151): python3-toml-0.10.2-6.el9.0.1.noarch 723 kB/s | 44 kB 00:00 2026-04-17T12:18:29.947 INFO:teuthology.orchestra.run.vm06.stdout:(139/151): qatlib-24.09.0-1.el9.x86_64.rpm 6.6 MB/s | 221 kB 00:00 2026-04-17T12:18:29.976 INFO:teuthology.orchestra.run.vm06.stdout:(140/151): qatlib-service-24.09.0-1.el9.x86_64. 1.2 MB/s | 36 kB 00:00 2026-04-17T12:18:29.982 INFO:teuthology.orchestra.run.vm08.stdout:(41/151): libarrow-9.0.0-15.el9.x86_64.rpm 37 MB/s | 4.4 MB 00:00 2026-04-17T12:18:29.986 INFO:teuthology.orchestra.run.vm08.stdout:(42/151): libarrow-doc-9.0.0-15.el9.noarch.rpm 5.6 MB/s | 25 kB 00:00 2026-04-17T12:18:29.991 INFO:teuthology.orchestra.run.vm08.stdout:(43/151): liboath-2.6.12-1.el9.x86_64.rpm 12 MB/s | 49 kB 00:00 2026-04-17T12:18:29.995 INFO:teuthology.orchestra.run.vm08.stdout:(44/151): libunwind-1.6.2-1.el9.x86_64.rpm 17 MB/s | 67 kB 00:00 2026-04-17T12:18:30.001 INFO:teuthology.orchestra.run.vm08.stdout:(45/151): luarocks-3.9.2-5.el9.noarch.rpm 25 MB/s | 151 kB 00:00 2026-04-17T12:18:30.005 INFO:teuthology.orchestra.run.vm06.stdout:(141/151): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.2 MB/s | 65 kB 00:00 2026-04-17T12:18:30.023 INFO:teuthology.orchestra.run.vm08.stdout:(46/151): parquet-libs-9.0.0-15.el9.x86_64.rpm 38 MB/s | 838 kB 00:00 2026-04-17T12:18:30.043 INFO:teuthology.orchestra.run.vm06.stdout:(142/151): socat-1.7.4.1-8.el9.x86_64.rpm 7.8 MB/s | 299 kB 00:00 2026-04-17T12:18:30.043 INFO:teuthology.orchestra.run.vm08.stdout:(47/151): python3-asyncssh-2.13.2-5.el9.noarch. 28 MB/s | 548 kB 00:00 2026-04-17T12:18:30.047 INFO:teuthology.orchestra.run.vm08.stdout:(48/151): python3-autocommand-2.2.2-8.el9.noarc 7.3 MB/s | 29 kB 00:00 2026-04-17T12:18:30.052 INFO:teuthology.orchestra.run.vm08.stdout:(49/151): python3-backports-tarfile-1.2.0-1.el9 13 MB/s | 60 kB 00:00 2026-04-17T12:18:30.056 INFO:teuthology.orchestra.run.vm08.stdout:(50/151): python3-bcrypt-3.2.2-1.el9.x86_64.rpm 11 MB/s | 43 kB 00:00 2026-04-17T12:18:30.061 INFO:teuthology.orchestra.run.vm08.stdout:(51/151): python3-cachetools-4.2.4-1.el9.noarch 7.2 MB/s | 32 kB 00:00 2026-04-17T12:18:30.064 INFO:teuthology.orchestra.run.vm08.stdout:(52/151): python3-certifi-2023.05.07-4.el9.noar 4.0 MB/s | 14 kB 00:00 2026-04-17T12:18:30.071 INFO:teuthology.orchestra.run.vm08.stdout:(53/151): python3-cheroot-10.0.1-5.el9.noarch.r 27 MB/s | 173 kB 00:00 2026-04-17T12:18:30.075 INFO:teuthology.orchestra.run.vm06.stdout:(143/151): xmlsec1-1.2.29-13.el9.x86_64.rpm 5.8 MB/s | 188 kB 00:00 2026-04-17T12:18:30.082 INFO:teuthology.orchestra.run.vm08.stdout:(54/151): python3-cherrypy-18.10.0-5.el9.noarch 25 MB/s | 290 kB 00:00 2026-04-17T12:18:30.091 INFO:teuthology.orchestra.run.vm08.stdout:(55/151): python3-google-auth-2.45.0-1.el9.noar 31 MB/s | 254 kB 00:00 2026-04-17T12:18:30.105 INFO:teuthology.orchestra.run.vm06.stdout:(144/151): xmlsec1-openssl-1.2.29-13.el9.x86_64 3.0 MB/s | 89 kB 00:00 2026-04-17T12:18:30.134 INFO:teuthology.orchestra.run.vm06.stdout:(145/151): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.1 MB/s | 63 kB 00:00 2026-04-17T12:18:30.140 INFO:teuthology.orchestra.run.vm08.stdout:(56/151): python3-grpcio-1.46.7-10.el9.x86_64.r 42 MB/s | 2.0 MB 00:00 2026-04-17T12:18:30.146 INFO:teuthology.orchestra.run.vm08.stdout:(57/151): python3-grpcio-tools-1.46.7-10.el9.x8 22 MB/s | 144 kB 00:00 2026-04-17T12:18:30.154 INFO:teuthology.orchestra.run.vm08.stdout:(58/151): python3-influxdb-5.3.1-1.el9.noarch.r 20 MB/s | 139 kB 00:00 2026-04-17T12:18:30.158 INFO:teuthology.orchestra.run.vm08.stdout:(59/151): python3-isodate-0.6.1-3.el9.noarch.rp 12 MB/s | 56 kB 00:00 2026-04-17T12:18:30.161 INFO:teuthology.orchestra.run.vm06.stdout:(146/151): lua-devel-5.4.4-4.el9.x86_64.rpm 786 kB/s | 21 kB 00:00 2026-04-17T12:18:30.161 INFO:teuthology.orchestra.run.vm08.stdout:(60/151): python3-jaraco-8.2.1-3.el9.noarch.rpm 3.9 MB/s | 11 kB 00:00 2026-04-17T12:18:30.166 INFO:teuthology.orchestra.run.vm08.stdout:(61/151): python3-jaraco-classes-3.2.1-5.el9.no 3.6 MB/s | 18 kB 00:00 2026-04-17T12:18:30.169 INFO:teuthology.orchestra.run.vm08.stdout:(62/151): python3-jaraco-collections-3.0.0-8.el 7.6 MB/s | 23 kB 00:00 2026-04-17T12:18:30.172 INFO:teuthology.orchestra.run.vm08.stdout:(63/151): python3-jaraco-context-6.0.1-3.el9.no 6.5 MB/s | 20 kB 00:00 2026-04-17T12:18:30.176 INFO:teuthology.orchestra.run.vm08.stdout:(64/151): python3-jaraco-functools-3.5.0-2.el9. 6.0 MB/s | 19 kB 00:00 2026-04-17T12:18:30.181 INFO:teuthology.orchestra.run.vm08.stdout:(65/151): python3-jaraco-text-4.0.0-2.el9.noarc 4.6 MB/s | 26 kB 00:00 2026-04-17T12:18:30.220 INFO:teuthology.orchestra.run.vm08.stdout:(66/151): python3-kubernetes-26.1.0-3.el9.noarc 27 MB/s | 1.0 MB 00:00 2026-04-17T12:18:30.225 INFO:teuthology.orchestra.run.vm06.stdout:(147/151): protobuf-compiler-3.14.0-17.el9_7.x8 13 MB/s | 862 kB 00:00 2026-04-17T12:18:30.225 INFO:teuthology.orchestra.run.vm08.stdout:(67/151): python3-more-itertools-8.12.0-2.el9.n 16 MB/s | 79 kB 00:00 2026-04-17T12:18:30.229 INFO:teuthology.orchestra.run.vm08.stdout:(68/151): python3-msgpack-1.0.3-2.el9.x86_64.rp 19 MB/s | 86 kB 00:00 2026-04-17T12:18:30.237 INFO:teuthology.orchestra.run.vm08.stdout:(69/151): python3-natsort-7.1.1-5.el9.noarch.rp 8.2 MB/s | 58 kB 00:00 2026-04-17T12:18:30.240 INFO:teuthology.orchestra.run.vm08.stdout:(70/151): python3-portend-3.1.0-2.el9.noarch.rp 4.9 MB/s | 16 kB 00:00 2026-04-17T12:18:30.247 INFO:teuthology.orchestra.run.vm08.stdout:(71/151): python3-pyOpenSSL-21.0.0-1.el9.noarch 13 MB/s | 90 kB 00:00 2026-04-17T12:18:30.250 INFO:teuthology.orchestra.run.vm08.stdout:(72/151): python3-repoze-lru-0.7-16.el9.noarch. 10 MB/s | 31 kB 00:00 2026-04-17T12:18:30.257 INFO:teuthology.orchestra.run.vm08.stdout:(73/151): python3-routes-2.5.1-5.el9.noarch.rpm 26 MB/s | 188 kB 00:00 2026-04-17T12:18:30.262 INFO:teuthology.orchestra.run.vm08.stdout:(74/151): python3-rsa-4.9-2.el9.noarch.rpm 13 MB/s | 59 kB 00:00 2026-04-17T12:18:30.268 INFO:teuthology.orchestra.run.vm08.stdout:(75/151): python3-saml-1.16.0-1.el9.noarch.rpm 23 MB/s | 125 kB 00:00 2026-04-17T12:18:30.272 INFO:teuthology.orchestra.run.vm08.stdout:(76/151): python3-tempora-5.0.0-2.el9.noarch.rp 9.9 MB/s | 36 kB 00:00 2026-04-17T12:18:30.278 INFO:teuthology.orchestra.run.vm08.stdout:(77/151): python3-typing-extensions-4.15.0-1.el 13 MB/s | 86 kB 00:00 2026-04-17T12:18:30.284 INFO:teuthology.orchestra.run.vm08.stdout:(78/151): python3-websocket-client-1.2.3-2.el9. 17 MB/s | 90 kB 00:00 2026-04-17T12:18:30.289 INFO:teuthology.orchestra.run.vm08.stdout:(79/151): python3-xmlsec-1.3.13-1.el9.x86_64.rp 12 MB/s | 48 kB 00:00 2026-04-17T12:18:30.293 INFO:teuthology.orchestra.run.vm08.stdout:(80/151): python3-xmltodict-0.12.0-15.el9.noarc 5.4 MB/s | 22 kB 00:00 2026-04-17T12:18:30.297 INFO:teuthology.orchestra.run.vm08.stdout:(81/151): python3-zc-lockfile-2.0-10.el9.noarch 5.9 MB/s | 20 kB 00:00 2026-04-17T12:18:30.304 INFO:teuthology.orchestra.run.vm08.stdout:(82/151): re2-20211101-20.el9.x86_64.rpm 30 MB/s | 191 kB 00:00 2026-04-17T12:18:30.310 INFO:teuthology.orchestra.run.vm08.stdout:(83/151): s3cmd-2.4.0-1.el9.noarch.rpm 34 MB/s | 206 kB 00:00 2026-04-17T12:18:30.370 INFO:teuthology.orchestra.run.vm08.stdout:(84/151): thrift-0.15.0-4.el9.x86_64.rpm 27 MB/s | 1.6 MB 00:00 2026-04-17T12:18:30.515 INFO:teuthology.orchestra.run.vm08.stdout:(85/151): bzip2-1.0.8-10.el9_5.x86_64.rpm 355 kB/s | 51 kB 00:00 2026-04-17T12:18:30.580 INFO:teuthology.orchestra.run.vm08.stdout:(86/151): c-ares-1.19.1-2.el9_4.x86_64.rpm 1.7 MB/s | 110 kB 00:00 2026-04-17T12:18:30.662 INFO:teuthology.orchestra.run.vm08.stdout:(87/151): cryptsetup-2.7.2-4.el9.x86_64.rpm 3.7 MB/s | 310 kB 00:00 2026-04-17T12:18:30.691 INFO:teuthology.orchestra.run.vm08.stdout:(88/151): fuse-2.9.9-17.el9.x86_64.rpm 2.6 MB/s | 78 kB 00:00 2026-04-17T12:18:30.718 INFO:teuthology.orchestra.run.vm08.stdout:(89/151): ledmon-libs-1.1.0-3.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-17T12:18:30.749 INFO:teuthology.orchestra.run.vm08.stdout:(90/151): libconfig-1.7.2-9.el9.x86_64.rpm 2.3 MB/s | 71 kB 00:00 2026-04-17T12:18:30.822 INFO:teuthology.orchestra.run.vm08.stdout:(91/151): libgfortran-11.5.0-11.el9.x86_64.rpm 11 MB/s | 794 kB 00:00 2026-04-17T12:18:30.822 INFO:teuthology.orchestra.run.vm06.stdout:(148/151): librados2-20.2.0-21.gc03ba9ecf58.el9 5.9 MB/s | 3.5 MB 00:00 2026-04-17T12:18:30.868 INFO:teuthology.orchestra.run.vm08.stdout:(92/151): libquadmath-11.5.0-11.el9.x86_64.rpm 4.0 MB/s | 184 kB 00:00 2026-04-17T12:18:30.869 INFO:teuthology.orchestra.run.vm06.stdout:(149/151): python3-scipy-1.9.3-2.el9.x86_64.rpm 18 MB/s | 19 MB 00:01 2026-04-17T12:18:30.917 INFO:teuthology.orchestra.run.vm08.stdout:(93/151): mailcap-2.1.49-5.el9.0.2.noarch.rpm 667 kB/s | 32 kB 00:00 2026-04-17T12:18:30.973 INFO:teuthology.orchestra.run.vm08.stdout:(94/151): nvme-cli-2.13-1.el9.x86_64.rpm 18 MB/s | 1.0 MB 00:00 2026-04-17T12:18:31.023 INFO:teuthology.orchestra.run.vm08.stdout:(95/151): pciutils-3.7.0-7.el9.x86_64.rpm 1.8 MB/s | 92 kB 00:00 2026-04-17T12:18:31.024 INFO:teuthology.orchestra.run.vm02.stdout:(151/151): ceph-test-20.2.0-21.gc03ba9ecf58.el9 9.5 MB/s | 85 MB 00:08 2026-04-17T12:18:31.029 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-04-17T12:18:31.029 INFO:teuthology.orchestra.run.vm02.stdout:Total 22 MB/s | 274 MB 00:12 2026-04-17T12:18:31.053 INFO:teuthology.orchestra.run.vm08.stdout:(96/151): python3-cffi-1.14.5-5.el9.x86_64.rpm 8.1 MB/s | 241 kB 00:00 2026-04-17T12:18:31.109 INFO:teuthology.orchestra.run.vm08.stdout:(97/151): python3-cryptography-36.0.1-5.el9_6.x 21 MB/s | 1.2 MB 00:00 2026-04-17T12:18:31.137 INFO:teuthology.orchestra.run.vm08.stdout:(98/151): python3-ply-3.11-14.el9.0.1.noarch.rp 3.6 MB/s | 103 kB 00:00 2026-04-17T12:18:31.168 INFO:teuthology.orchestra.run.vm08.stdout:(99/151): python3-pycparser-2.20-6.el9.noarch.r 4.0 MB/s | 124 kB 00:00 2026-04-17T12:18:31.198 INFO:teuthology.orchestra.run.vm08.stdout:(100/151): python3-pyparsing-2.4.7-9.el9.0.1.no 5.0 MB/s | 150 kB 00:00 2026-04-17T12:18:31.230 INFO:teuthology.orchestra.run.vm08.stdout:(101/151): python3-requests-2.25.1-10.el9_6.noa 3.6 MB/s | 115 kB 00:00 2026-04-17T12:18:31.247 INFO:teuthology.orchestra.run.vm06.stdout:(150/151): librbd1-20.2.0-21.gc03ba9ecf58.el9.c 6.7 MB/s | 2.8 MB 00:00 2026-04-17T12:18:31.258 INFO:teuthology.orchestra.run.vm08.stdout:(102/151): python3-urllib3-1.26.5-6.el9_7.1.noa 6.7 MB/s | 191 kB 00:00 2026-04-17T12:18:31.292 INFO:teuthology.orchestra.run.vm08.stdout:(103/151): smartmontools-7.2-9.el9.x86_64.rpm 16 MB/s | 551 kB 00:00 2026-04-17T12:18:31.322 INFO:teuthology.orchestra.run.vm08.stdout:(104/151): unzip-6.0-59.el9.x86_64.rpm 6.0 MB/s | 180 kB 00:00 2026-04-17T12:18:31.351 INFO:teuthology.orchestra.run.vm08.stdout:(105/151): zip-3.0-35.el9.x86_64.rpm 8.8 MB/s | 263 kB 00:00 2026-04-17T12:18:31.384 INFO:teuthology.orchestra.run.vm08.stdout:(106/151): boost-program-options-1.75.0-13.el9_ 3.1 MB/s | 104 kB 00:00 2026-04-17T12:18:31.411 INFO:teuthology.orchestra.run.vm08.stdout:(107/151): flexiblas-3.0.4-8.el9.0.1.x86_64.rpm 1.1 MB/s | 30 kB 00:00 2026-04-17T12:18:31.527 INFO:teuthology.orchestra.run.vm08.stdout:(108/151): flexiblas-netlib-3.0.4-8.el9.0.1.x86 26 MB/s | 3.0 MB 00:00 2026-04-17T12:18:31.554 INFO:teuthology.orchestra.run.vm08.stdout:(109/151): flexiblas-openblas-openmp-3.0.4-8.el 570 kB/s | 15 kB 00:00 2026-04-17T12:18:31.583 INFO:teuthology.orchestra.run.vm08.stdout:(110/151): libnbd-1.20.3-4.el9.x86_64.rpm 5.8 MB/s | 171 kB 00:00 2026-04-17T12:18:31.612 INFO:teuthology.orchestra.run.vm08.stdout:(111/151): libpmemobj-1.12.1-1.el9.x86_64.rpm 5.6 MB/s | 159 kB 00:00 2026-04-17T12:18:31.639 INFO:teuthology.orchestra.run.vm08.stdout:(112/151): librabbitmq-0.11.0-7.el9.x86_64.rpm 1.6 MB/s | 44 kB 00:00 2026-04-17T12:18:31.674 INFO:teuthology.orchestra.run.vm08.stdout:(113/151): librdkafka-1.6.1-102.el9.x86_64.rpm 19 MB/s | 662 kB 00:00 2026-04-17T12:18:31.707 INFO:teuthology.orchestra.run.vm08.stdout:(114/151): libstoragemgmt-1.10.1-1.el9.x86_64.r 7.2 MB/s | 243 kB 00:00 2026-04-17T12:18:31.737 INFO:teuthology.orchestra.run.vm08.stdout:(115/151): libxslt-1.1.34-14.el9_7.1.x86_64.rpm 8.0 MB/s | 240 kB 00:00 2026-04-17T12:18:31.769 INFO:teuthology.orchestra.run.vm08.stdout:(116/151): lttng-ust-2.12.0-6.el9.x86_64.rpm 8.7 MB/s | 282 kB 00:00 2026-04-17T12:18:31.799 INFO:teuthology.orchestra.run.vm08.stdout:(117/151): lua-5.4.4-4.el9.x86_64.rpm 6.1 MB/s | 187 kB 00:00 2026-04-17T12:18:31.827 INFO:teuthology.orchestra.run.vm08.stdout:(118/151): openblas-0.3.29-1.el9.x86_64.rpm 1.5 MB/s | 41 kB 00:00 2026-04-17T12:18:31.950 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-04-17T12:18:31.996 INFO:teuthology.orchestra.run.vm08.stdout:(119/151): openblas-openmp-0.3.29-1.el9.x86_64. 31 MB/s | 5.3 MB 00:00 2026-04-17T12:18:32.018 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-04-17T12:18:32.018 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-04-17T12:18:32.025 INFO:teuthology.orchestra.run.vm08.stdout:(120/151): perl-Benchmark-1.23-481.1.el9_6.noar 896 kB/s | 25 kB 00:00 2026-04-17T12:18:32.055 INFO:teuthology.orchestra.run.vm08.stdout:(121/151): perl-Test-Harness-3.42-461.el9.noarc 9.0 MB/s | 267 kB 00:00 2026-04-17T12:18:32.097 INFO:teuthology.orchestra.run.vm08.stdout:(122/151): protobuf-3.14.0-17.el9_7.x86_64.rpm 24 MB/s | 1.0 MB 00:00 2026-04-17T12:18:32.253 INFO:teuthology.orchestra.run.vm08.stdout:(123/151): python3-babel-2.9.1-2.el9.noarch.rpm 37 MB/s | 5.8 MB 00:00 2026-04-17T12:18:32.283 INFO:teuthology.orchestra.run.vm08.stdout:(124/151): python3-devel-3.9.23-2.el9.x86_64.rp 6.8 MB/s | 205 kB 00:00 2026-04-17T12:18:32.314 INFO:teuthology.orchestra.run.vm08.stdout:(125/151): python3-jinja2-2.11.3-8.el9_5.noarch 7.2 MB/s | 228 kB 00:00 2026-04-17T12:18:32.344 INFO:teuthology.orchestra.run.vm08.stdout:(126/151): python3-jmespath-1.0.1-1.el9_7.noarc 1.4 MB/s | 43 kB 00:00 2026-04-17T12:18:32.372 INFO:teuthology.orchestra.run.vm08.stdout:(127/151): python3-libstoragemgmt-1.10.1-1.el9. 5.8 MB/s | 166 kB 00:00 2026-04-17T12:18:32.445 INFO:teuthology.orchestra.run.vm08.stdout:(128/151): python3-lxml-4.6.5-3.el9.x86_64.rpm 16 MB/s | 1.2 MB 00:00 2026-04-17T12:18:32.512 INFO:teuthology.orchestra.run.vm08.stdout:(129/151): python3-markupsafe-1.1.1-12.el9.x86_ 477 kB/s | 32 kB 00:00 2026-04-17T12:18:32.801 INFO:teuthology.orchestra.run.vm08.stdout:(130/151): python3-numpy-1.23.5-2.el9_7.x86_64. 20 MB/s | 5.8 MB 00:00 2026-04-17T12:18:32.833 INFO:teuthology.orchestra.run.vm08.stdout:(131/151): python3-numpy-f2py-1.23.5-2.el9_7.x8 11 MB/s | 368 kB 00:00 2026-04-17T12:18:32.865 INFO:teuthology.orchestra.run.vm08.stdout:(132/151): python3-packaging-20.9-5.el9.noarch. 2.1 MB/s | 69 kB 00:00 2026-04-17T12:18:32.895 INFO:teuthology.orchestra.run.vm08.stdout:(133/151): python3-protobuf-3.14.0-17.el9_7.noa 7.9 MB/s | 237 kB 00:00 2026-04-17T12:18:32.923 INFO:teuthology.orchestra.run.vm08.stdout:(134/151): python3-pyasn1-0.4.8-7.el9_7.noarch. 4.7 MB/s | 132 kB 00:00 2026-04-17T12:18:32.952 INFO:teuthology.orchestra.run.vm08.stdout:(135/151): python3-pyasn1-modules-0.4.8-7.el9_7 7.1 MB/s | 210 kB 00:00 2026-04-17T12:18:32.979 INFO:teuthology.orchestra.run.vm08.stdout:(136/151): python3-requests-oauthlib-1.3.0-12.e 1.6 MB/s | 43 kB 00:00 2026-04-17T12:18:33.111 INFO:teuthology.orchestra.run.vm08.stdout:(137/151): ceph-mgr-dashboard-20.2.0-21.gc03ba9 2.9 MB/s | 15 MB 00:05 2026-04-17T12:18:33.138 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-04-17T12:18:33.138 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-04-17T12:18:33.246 INFO:teuthology.orchestra.run.vm08.stdout:(138/151): python3-toml-0.10.2-6.el9.0.1.noarch 328 kB/s | 44 kB 00:00 2026-04-17T12:18:33.365 INFO:teuthology.orchestra.run.vm08.stdout:(139/151): python3-scipy-1.9.3-2.el9.x86_64.rpm 49 MB/s | 19 MB 00:00 2026-04-17T12:18:33.367 INFO:teuthology.orchestra.run.vm08.stdout:(140/151): qatlib-24.09.0-1.el9.x86_64.rpm 1.8 MB/s | 221 kB 00:00 2026-04-17T12:18:33.393 INFO:teuthology.orchestra.run.vm08.stdout:(141/151): qatlib-service-24.09.0-1.el9.x86_64. 1.3 MB/s | 36 kB 00:00 2026-04-17T12:18:33.394 INFO:teuthology.orchestra.run.vm08.stdout:(142/151): qatzip-libs-1.3.1-1.el9.x86_64.rpm 2.4 MB/s | 65 kB 00:00 2026-04-17T12:18:33.442 INFO:teuthology.orchestra.run.vm08.stdout:(143/151): socat-1.7.4.1-8.el9.x86_64.rpm 6.0 MB/s | 299 kB 00:00 2026-04-17T12:18:33.444 INFO:teuthology.orchestra.run.vm08.stdout:(144/151): xmlsec1-1.2.29-13.el9.x86_64.rpm 3.7 MB/s | 188 kB 00:00 2026-04-17T12:18:33.470 INFO:teuthology.orchestra.run.vm08.stdout:(145/151): xmlsec1-openssl-1.2.29-13.el9.x86_64 3.2 MB/s | 89 kB 00:00 2026-04-17T12:18:33.471 INFO:teuthology.orchestra.run.vm08.stdout:(146/151): xmlstarlet-1.6.1-20.el9.x86_64.rpm 2.3 MB/s | 63 kB 00:00 2026-04-17T12:18:33.497 INFO:teuthology.orchestra.run.vm08.stdout:(147/151): lua-devel-5.4.4-4.el9.x86_64.rpm 792 kB/s | 21 kB 00:00 2026-04-17T12:18:33.531 INFO:teuthology.orchestra.run.vm08.stdout:(148/151): protobuf-compiler-3.14.0-17.el9_7.x8 14 MB/s | 862 kB 00:00 2026-04-17T12:18:34.331 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-04-17T12:18:34.341 INFO:teuthology.orchestra.run.vm02.stdout: Installing : thrift-0.15.0-4.el9.x86_64 1/153 2026-04-17T12:18:34.344 INFO:teuthology.orchestra.run.vm08.stdout:(149/151): librados2-20.2.0-21.gc03ba9ecf58.el9 4.1 MB/s | 3.5 MB 00:00 2026-04-17T12:18:34.356 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 2/153 2026-04-17T12:18:34.562 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/153 2026-04-17T12:18:34.566 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:18:34.630 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:18:34.633 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 5/153 2026-04-17T12:18:34.652 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 5/153 2026-04-17T12:18:34.656 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/153 2026-04-17T12:18:34.659 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 7/153 2026-04-17T12:18:34.697 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 7/153 2026-04-17T12:18:34.707 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso 8/153 2026-04-17T12:18:34.722 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libxslt-1.1.34-14.el9_7.1.x86_64 9/153 2026-04-17T12:18:34.727 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 10/153 2026-04-17T12:18:34.734 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 11/153 2026-04-17T12:18:34.738 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 12/153 2026-04-17T12:18:34.745 INFO:teuthology.orchestra.run.vm02.stdout: Installing : liboath-2.6.12-1.el9.x86_64 13/153 2026-04-17T12:18:34.911 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 14/153 2026-04-17T12:18:34.914 INFO:teuthology.orchestra.run.vm02.stdout: Upgrading : librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 15/153 2026-04-17T12:18:34.931 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 15/153 2026-04-17T12:18:34.974 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 16/153 2026-04-17T12:18:34.983 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 17/153 2026-04-17T12:18:34.990 INFO:teuthology.orchestra.run.vm08.stdout:(150/151): librbd1-20.2.0-21.gc03ba9ecf58.el9.c 2.0 MB/s | 2.8 MB 00:01 2026-04-17T12:18:34.994 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 18/153 2026-04-17T12:18:34.996 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 19/153 2026-04-17T12:18:35.029 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 19/153 2026-04-17T12:18:35.032 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 20/153 2026-04-17T12:18:35.077 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 20/153 2026-04-17T12:18:35.090 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-0.4.8-7.el9_7.noarch 21/153 2026-04-17T12:18:35.098 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-3.14.0-17.el9_7.x86_64 22/153 2026-04-17T12:18:35.103 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-5.4.4-4.el9.x86_64 23/153 2026-04-17T12:18:35.111 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-3.0.4-8.el9.0.1.x86_64 24/153 2026-04-17T12:18:35.143 INFO:teuthology.orchestra.run.vm02.stdout: Installing : unzip-6.0-59.el9.x86_64 25/153 2026-04-17T12:18:35.164 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-urllib3-1.26.5-6.el9_7.1.noarch 26/153 2026-04-17T12:18:35.170 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-2.25.1-10.el9_6.noarch 27/153 2026-04-17T12:18:35.178 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libquadmath-11.5.0-11.el9.x86_64 28/153 2026-04-17T12:18:35.182 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libgfortran-11.5.0-11.el9.x86_64 29/153 2026-04-17T12:18:35.188 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 30/153 2026-04-17T12:18:35.227 INFO:teuthology.orchestra.run.vm02.stdout: Installing : re2-1:20211101-20.el9.x86_64 31/153 2026-04-17T12:18:35.266 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 32/153 2026-04-17T12:18:35.276 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 33/153 2026-04-17T12:18:35.297 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.e 34/153 2026-04-17T12:18:35.315 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clys 35/153 2026-04-17T12:18:35.327 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 36/153 2026-04-17T12:18:35.362 INFO:teuthology.orchestra.run.vm02.stdout: Installing : zip-3.0-35.el9.x86_64 37/153 2026-04-17T12:18:35.371 INFO:teuthology.orchestra.run.vm02.stdout: Installing : luarocks-3.9.2-5.el9.noarch 38/153 2026-04-17T12:18:35.384 INFO:teuthology.orchestra.run.vm02.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 39/153 2026-04-17T12:18:35.410 INFO:teuthology.orchestra.run.vm02.stdout: Installing : protobuf-compiler-3.14.0-17.el9_7.x86_64 40/153 2026-04-17T12:18:35.483 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rsa-4.9-2.el9.noarch 41/153 2026-04-17T12:18:35.489 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 42/153 2026-04-17T12:18:35.497 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/153 2026-04-17T12:18:35.507 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/153 2026-04-17T12:18:35.529 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/153 2026-04-17T12:18:35.536 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 46/153 2026-04-17T12:18:35.548 INFO:teuthology.orchestra.run.vm02.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 47/153 2026-04-17T12:18:35.560 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clys 48/153 2026-04-17T12:18:35.571 INFO:teuthology.orchestra.run.vm02.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 49/153 2026-04-17T12:18:35.578 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-toml-0.10.2-6.el9.0.1.noarch 50/153 2026-04-17T12:18:35.589 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 51/153 2026-04-17T12:18:35.596 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 52/153 2026-04-17T12:18:35.641 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 53/153 2026-04-17T12:18:35.659 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-protobuf-3.14.0-17.el9_7.noarch 54/153 2026-04-17T12:18:35.671 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 55/153 2026-04-17T12:18:35.723 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jmespath-1.0.1-1.el9_7.noarch 56/153 2026-04-17T12:18:36.004 INFO:teuthology.orchestra.run.vm06.stdout:(151/151): ceph-test-20.2.0-21.gc03ba9ecf58.el9 6.2 MB/s | 85 MB 00:13 2026-04-17T12:18:36.010 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-04-17T12:18:36.010 INFO:teuthology.orchestra.run.vm06.stdout:Total 16 MB/s | 274 MB 00:16 2026-04-17T12:18:36.059 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-devel-3.9.23-2.el9.x86_64 57/153 2026-04-17T12:18:36.095 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 58/153 2026-04-17T12:18:36.101 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jinja2-2.11.3-8.el9_5.noarch 59/153 2026-04-17T12:18:36.107 INFO:teuthology.orchestra.run.vm02.stdout: Installing : perl-Benchmark-1.23-481.1.el9_6.noarch 60/153 2026-04-17T12:18:36.181 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-0.3.29-1.el9.x86_64 61/153 2026-04-17T12:18:36.186 INFO:teuthology.orchestra.run.vm02.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 62/153 2026-04-17T12:18:36.218 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 63/153 2026-04-17T12:18:36.684 INFO:teuthology.orchestra.run.vm02.stdout: Installing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 64/153 2026-04-17T12:18:36.788 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-1:1.23.5-2.el9_7.x86_64 65/153 2026-04-17T12:18:37.017 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-04-17T12:18:37.083 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-04-17T12:18:37.083 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-04-17T12:18:37.804 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 66/153 2026-04-17T12:18:37.834 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 67/153 2026-04-17T12:18:37.859 INFO:teuthology.orchestra.run.vm02.stdout: Installing : boost-program-options-1.75.0-13.el9_7.x86_64 68/153 2026-04-17T12:18:37.861 INFO:teuthology.orchestra.run.vm02.stdout: Installing : smartmontools-1:7.2-9.el9.x86_64 69/153 2026-04-17T12:18:37.881 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: smartmontools-1:7.2-9.el9.x86_64 69/153 2026-04-17T12:18:37.881 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-04-17T12:18:37.881 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:37.909 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 70/153 2026-04-17T12:18:37.923 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-packaging-20.9-5.el9.noarch 71/153 2026-04-17T12:18:37.948 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-ply-3.11-14.el9.0.1.noarch 72/153 2026-04-17T12:18:37.979 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 73/153 2026-04-17T12:18:38.127 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 74/153 2026-04-17T12:18:38.144 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cryptography-36.0.1-5.el9_6.x86_64 75/153 2026-04-17T12:18:38.176 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 76/153 2026-04-17T12:18:38.187 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cheroot-10.0.1-5.el9.noarch 77/153 2026-04-17T12:18:38.194 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 78/153 2026-04-17T12:18:38.197 INFO:teuthology.orchestra.run.vm02.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 79/153 2026-04-17T12:18:38.219 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-04-17T12:18:38.220 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-04-17T12:18:38.238 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-24.09.0-1.el9.x86_64 80/153 2026-04-17T12:18:38.245 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-24.09.0-1.el9.x86_64 80/153 2026-04-17T12:18:38.248 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatlib-service-24.09.0-1.el9.x86_64 81/153 2026-04-17T12:18:38.279 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: qatlib-service-24.09.0-1.el9.x86_64 81/153 2026-04-17T12:18:38.451 INFO:teuthology.orchestra.run.vm02.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 82/153 2026-04-17T12:18:38.458 INFO:teuthology.orchestra.run.vm02.stdout: Installing : nvme-cli-2.13-1.el9.x86_64 83/153 2026-04-17T12:18:38.869 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: nvme-cli-2.13-1.el9.x86_64 83/153 2026-04-17T12:18:38.869 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-04-17T12:18:38.869 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:38.879 INFO:teuthology.orchestra.run.vm02.stdout: Installing : mailcap-2.1.49-5.el9.0.2.noarch 84/153 2026-04-17T12:18:38.886 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 85/153 2026-04-17T12:18:38.918 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:38.918 INFO:teuthology.orchestra.run.vm02.stdout:Creating group 'libstoragemgmt' with GID 992. 2026-04-17T12:18:38.918 INFO:teuthology.orchestra.run.vm02.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 992 and GID 992. 2026-04-17T12:18:38.918 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:38.939 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:38.979 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:38.979 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-04-17T12:18:38.979 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:39.022 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 87/153 2026-04-17T12:18:39.062 INFO:teuthology.orchestra.run.vm02.stdout: Installing : fuse-2.9.9-17.el9.x86_64 88/153 2026-04-17T12:18:39.106 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cryptsetup-2.7.2-4.el9.x86_64 89/153 2026-04-17T12:18:39.128 INFO:teuthology.orchestra.run.vm02.stdout: Installing : c-ares-1.19.1-2.el9_4.x86_64 90/153 2026-04-17T12:18:39.147 INFO:teuthology.orchestra.run.vm08.stdout:(151/151): ceph-test-20.2.0-21.gc03ba9ecf58.el9 5.4 MB/s | 85 MB 00:15 2026-04-17T12:18:39.153 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 91/153 2026-04-17T12:18:39.154 INFO:teuthology.orchestra.run.vm08.stdout:-------------------------------------------------------------------------------- 2026-04-17T12:18:39.154 INFO:teuthology.orchestra.run.vm08.stdout:Total 13 MB/s | 274 MB 00:20 2026-04-17T12:18:39.178 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 92/153 2026-04-17T12:18:39.188 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 93/153 2026-04-17T12:18:39.269 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 94/153 2026-04-17T12:18:39.285 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 95/153 2026-04-17T12:18:39.305 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 96/153 2026-04-17T12:18:39.323 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 97/153 2026-04-17T12:18:39.336 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 98/153 2026-04-17T12:18:39.370 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 99/153 2026-04-17T12:18:39.390 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 100/153 2026-04-17T12:18:39.423 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 101/153 2026-04-17T12:18:39.436 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 102/153 2026-04-17T12:18:39.456 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 103/153 2026-04-17T12:18:39.478 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-04-17T12:18:39.489 INFO:teuthology.orchestra.run.vm06.stdout: Installing : thrift-0.15.0-4.el9.x86_64 1/153 2026-04-17T12:18:39.504 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 2/153 2026-04-17T12:18:39.520 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 104/153 2026-04-17T12:18:39.707 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/153 2026-04-17T12:18:39.720 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:18:39.799 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:18:39.803 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 5/153 2026-04-17T12:18:39.829 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 5/153 2026-04-17T12:18:39.833 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/153 2026-04-17T12:18:39.836 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 7/153 2026-04-17T12:18:39.878 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 7/153 2026-04-17T12:18:39.887 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso 8/153 2026-04-17T12:18:39.902 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libxslt-1.1.34-14.el9_7.1.x86_64 9/153 2026-04-17T12:18:39.907 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 10/153 2026-04-17T12:18:39.913 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 11/153 2026-04-17T12:18:39.978 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 12/153 2026-04-17T12:18:39.988 INFO:teuthology.orchestra.run.vm06.stdout: Installing : liboath-2.6.12-1.el9.x86_64 13/153 2026-04-17T12:18:40.001 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-04-17T12:18:40.039 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 105/153 2026-04-17T12:18:40.068 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-04-17T12:18:40.068 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-04-17T12:18:40.075 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 106/153 2026-04-17T12:18:40.088 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 107/153 2026-04-17T12:18:40.101 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 108/153 2026-04-17T12:18:40.117 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 109/153 2026-04-17T12:18:40.126 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 110/153 2026-04-17T12:18:40.151 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 14/153 2026-04-17T12:18:40.154 INFO:teuthology.orchestra.run.vm06.stdout: Upgrading : librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 15/153 2026-04-17T12:18:40.172 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 15/153 2026-04-17T12:18:40.186 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 111/153 2026-04-17T12:18:40.197 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-cherrypy-18.10.0-5.el9.noarch 112/153 2026-04-17T12:18:40.210 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 113/153 2026-04-17T12:18:40.216 INFO:teuthology.orchestra.run.vm02.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 114/153 2026-04-17T12:18:40.222 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 16/153 2026-04-17T12:18:40.225 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 115/153 2026-04-17T12:18:40.235 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 17/153 2026-04-17T12:18:40.267 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 18/153 2026-04-17T12:18:40.269 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 19/153 2026-04-17T12:18:40.304 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 19/153 2026-04-17T12:18:40.307 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 20/153 2026-04-17T12:18:40.360 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 20/153 2026-04-17T12:18:40.378 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-0.4.8-7.el9_7.noarch 21/153 2026-04-17T12:18:40.388 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-3.14.0-17.el9_7.x86_64 22/153 2026-04-17T12:18:40.394 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-5.4.4-4.el9.x86_64 23/153 2026-04-17T12:18:40.402 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-3.0.4-8.el9.0.1.x86_64 24/153 2026-04-17T12:18:40.438 INFO:teuthology.orchestra.run.vm06.stdout: Installing : unzip-6.0-59.el9.x86_64 25/153 2026-04-17T12:18:40.460 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-6.el9_7.1.noarch 26/153 2026-04-17T12:18:40.469 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-2.25.1-10.el9_6.noarch 27/153 2026-04-17T12:18:40.480 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libquadmath-11.5.0-11.el9.x86_64 28/153 2026-04-17T12:18:40.484 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libgfortran-11.5.0-11.el9.x86_64 29/153 2026-04-17T12:18:40.491 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 30/153 2026-04-17T12:18:40.538 INFO:teuthology.orchestra.run.vm06.stdout: Installing : re2-1:20211101-20.el9.x86_64 31/153 2026-04-17T12:18:40.579 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 32/153 2026-04-17T12:18:40.588 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 33/153 2026-04-17T12:18:40.600 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.e 34/153 2026-04-17T12:18:40.618 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clys 35/153 2026-04-17T12:18:40.631 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 36/153 2026-04-17T12:18:40.639 INFO:teuthology.orchestra.run.vm02.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 116/153 2026-04-17T12:18:40.663 INFO:teuthology.orchestra.run.vm02.stdout: Installing : librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 117/153 2026-04-17T12:18:40.665 INFO:teuthology.orchestra.run.vm06.stdout: Installing : zip-3.0-35.el9.x86_64 37/153 2026-04-17T12:18:40.674 INFO:teuthology.orchestra.run.vm06.stdout: Installing : luarocks-3.9.2-5.el9.noarch 38/153 2026-04-17T12:18:40.687 INFO:teuthology.orchestra.run.vm06.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 39/153 2026-04-17T12:18:40.695 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 117/153 2026-04-17T12:18:40.698 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 118/153 2026-04-17T12:18:40.713 INFO:teuthology.orchestra.run.vm06.stdout: Installing : protobuf-compiler-3.14.0-17.el9_7.x86_64 40/153 2026-04-17T12:18:40.786 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rsa-4.9-2.el9.noarch 41/153 2026-04-17T12:18:40.793 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 42/153 2026-04-17T12:18:40.804 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/153 2026-04-17T12:18:40.818 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/153 2026-04-17T12:18:40.841 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/153 2026-04-17T12:18:40.849 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 46/153 2026-04-17T12:18:40.862 INFO:teuthology.orchestra.run.vm06.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 47/153 2026-04-17T12:18:40.877 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clys 48/153 2026-04-17T12:18:40.889 INFO:teuthology.orchestra.run.vm06.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 49/153 2026-04-17T12:18:40.896 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-toml-0.10.2-6.el9.0.1.noarch 50/153 2026-04-17T12:18:40.910 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 51/153 2026-04-17T12:18:40.917 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 52/153 2026-04-17T12:18:40.974 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 53/153 2026-04-17T12:18:40.987 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-protobuf-3.14.0-17.el9_7.noarch 54/153 2026-04-17T12:18:41.002 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 55/153 2026-04-17T12:18:41.050 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9_7.noarch 56/153 2026-04-17T12:18:41.233 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-04-17T12:18:41.233 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-04-17T12:18:41.405 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-devel-3.9.23-2.el9.x86_64 57/153 2026-04-17T12:18:41.443 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 58/153 2026-04-17T12:18:41.448 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jinja2-2.11.3-8.el9_5.noarch 59/153 2026-04-17T12:18:41.452 INFO:teuthology.orchestra.run.vm06.stdout: Installing : perl-Benchmark-1.23-481.1.el9_6.noarch 60/153 2026-04-17T12:18:41.529 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-0.3.29-1.el9.x86_64 61/153 2026-04-17T12:18:41.534 INFO:teuthology.orchestra.run.vm06.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 62/153 2026-04-17T12:18:41.572 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 63/153 2026-04-17T12:18:42.032 INFO:teuthology.orchestra.run.vm06.stdout: Installing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 64/153 2026-04-17T12:18:42.140 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:42.140 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-1:1.23.5-2.el9_7.x86_64 65/153 2026-04-17T12:18:42.146 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:42.474 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-04-17T12:18:42.487 INFO:teuthology.orchestra.run.vm08.stdout: Installing : thrift-0.15.0-4.el9.x86_64 1/153 2026-04-17T12:18:42.500 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-more-itertools-8.12.0-2.el9.noarch 2/153 2026-04-17T12:18:42.533 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:42.540 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 120/153 2026-04-17T12:18:42.596 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 120/153 2026-04-17T12:18:42.596 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-04-17T12:18:42.596 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-04-17T12:18:42.596 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:42.603 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 121/153 2026-04-17T12:18:42.720 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lttng-ust-2.12.0-6.el9.x86_64 3/153 2026-04-17T12:18:42.724 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:18:42.799 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:18:42.801 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 5/153 2026-04-17T12:18:42.823 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 5/153 2026-04-17T12:18:42.827 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/153 2026-04-17T12:18:42.829 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 7/153 2026-04-17T12:18:42.868 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 7/153 2026-04-17T12:18:42.877 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso 8/153 2026-04-17T12:18:42.890 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libxslt-1.1.34-14.el9_7.1.x86_64 9/153 2026-04-17T12:18:42.895 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librdkafka-1.6.1-102.el9.x86_64 10/153 2026-04-17T12:18:42.900 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librabbitmq-0.11.0-7.el9.x86_64 11/153 2026-04-17T12:18:42.904 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libpmemobj-1.12.1-1.el9.x86_64 12/153 2026-04-17T12:18:42.910 INFO:teuthology.orchestra.run.vm08.stdout: Installing : liboath-2.6.12-1.el9.x86_64 13/153 2026-04-17T12:18:43.079 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libnbd-1.20.3-4.el9.x86_64 14/153 2026-04-17T12:18:43.082 INFO:teuthology.orchestra.run.vm08.stdout: Upgrading : librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 15/153 2026-04-17T12:18:43.100 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 15/153 2026-04-17T12:18:43.138 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 66/153 2026-04-17T12:18:43.152 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-8.2.1-3.el9.noarch 16/153 2026-04-17T12:18:43.164 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 67/153 2026-04-17T12:18:43.165 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-lxml-4.6.5-3.el9.x86_64 17/153 2026-04-17T12:18:43.177 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlsec1-1.2.29-13.el9.x86_64 18/153 2026-04-17T12:18:43.179 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 19/153 2026-04-17T12:18:43.184 INFO:teuthology.orchestra.run.vm06.stdout: Installing : boost-program-options-1.75.0-13.el9_7.x86_64 68/153 2026-04-17T12:18:43.195 INFO:teuthology.orchestra.run.vm06.stdout: Installing : smartmontools-1:7.2-9.el9.x86_64 69/153 2026-04-17T12:18:43.212 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 19/153 2026-04-17T12:18:43.213 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 20/153 2026-04-17T12:18:43.217 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: smartmontools-1:7.2-9.el9.x86_64 69/153 2026-04-17T12:18:43.217 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-04-17T12:18:43.217 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:43.241 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 70/153 2026-04-17T12:18:43.253 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-packaging-20.9-5.el9.noarch 71/153 2026-04-17T12:18:43.257 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 20/153 2026-04-17T12:18:43.274 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-0.4.8-7.el9_7.noarch 21/153 2026-04-17T12:18:43.275 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-ply-3.11-14.el9.0.1.noarch 72/153 2026-04-17T12:18:43.285 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-3.14.0-17.el9_7.x86_64 22/153 2026-04-17T12:18:43.290 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-5.4.4-4.el9.x86_64 23/153 2026-04-17T12:18:43.297 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-3.0.4-8.el9.0.1.x86_64 24/153 2026-04-17T12:18:43.297 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 73/153 2026-04-17T12:18:43.331 INFO:teuthology.orchestra.run.vm08.stdout: Installing : unzip-6.0-59.el9.x86_64 25/153 2026-04-17T12:18:43.351 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-urllib3-1.26.5-6.el9_7.1.noarch 26/153 2026-04-17T12:18:43.356 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-2.25.1-10.el9_6.noarch 27/153 2026-04-17T12:18:43.366 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libquadmath-11.5.0-11.el9.x86_64 28/153 2026-04-17T12:18:43.369 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libgfortran-11.5.0-11.el9.x86_64 29/153 2026-04-17T12:18:43.376 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ledmon-libs-1.1.0-3.el9.x86_64 30/153 2026-04-17T12:18:43.409 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 74/153 2026-04-17T12:18:43.419 INFO:teuthology.orchestra.run.vm08.stdout: Installing : re2-1:20211101-20.el9.x86_64 31/153 2026-04-17T12:18:43.425 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cryptography-36.0.1-5.el9_6.x86_64 75/153 2026-04-17T12:18:43.459 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-9.0.0-15.el9.x86_64 32/153 2026-04-17T12:18:43.462 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 76/153 2026-04-17T12:18:43.468 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 33/153 2026-04-17T12:18:43.472 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cheroot-10.0.1-5.el9.noarch 77/153 2026-04-17T12:18:43.479 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 78/153 2026-04-17T12:18:43.480 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.e 34/153 2026-04-17T12:18:43.482 INFO:teuthology.orchestra.run.vm06.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 79/153 2026-04-17T12:18:43.496 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clys 35/153 2026-04-17T12:18:43.506 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-requests-oauthlib-1.3.0-12.el9.noarch 36/153 2026-04-17T12:18:43.521 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-24.09.0-1.el9.x86_64 80/153 2026-04-17T12:18:43.528 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-24.09.0-1.el9.x86_64 80/153 2026-04-17T12:18:43.530 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatlib-service-24.09.0-1.el9.x86_64 81/153 2026-04-17T12:18:43.541 INFO:teuthology.orchestra.run.vm08.stdout: Installing : zip-3.0-35.el9.x86_64 37/153 2026-04-17T12:18:43.547 INFO:teuthology.orchestra.run.vm08.stdout: Installing : luarocks-3.9.2-5.el9.noarch 38/153 2026-04-17T12:18:43.557 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: qatlib-service-24.09.0-1.el9.x86_64 81/153 2026-04-17T12:18:43.557 INFO:teuthology.orchestra.run.vm08.stdout: Installing : lua-devel-5.4.4-4.el9.x86_64 39/153 2026-04-17T12:18:43.579 INFO:teuthology.orchestra.run.vm08.stdout: Installing : protobuf-compiler-3.14.0-17.el9_7.x86_64 40/153 2026-04-17T12:18:43.646 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rsa-4.9-2.el9.noarch 41/153 2026-04-17T12:18:43.653 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 42/153 2026-04-17T12:18:43.661 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlsec1-openssl-1.2.29-13.el9.x86_64 43/153 2026-04-17T12:18:43.670 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmlsec-1.3.13-1.el9.x86_64 44/153 2026-04-17T12:18:43.693 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-classes-3.2.1-5.el9.noarch 45/153 2026-04-17T12:18:43.699 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 46/153 2026-04-17T12:18:43.711 INFO:teuthology.orchestra.run.vm08.stdout: Installing : xmlstarlet-1.6.1-20.el9.x86_64 47/153 2026-04-17T12:18:43.719 INFO:teuthology.orchestra.run.vm06.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 82/153 2026-04-17T12:18:43.724 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clys 48/153 2026-04-17T12:18:43.725 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.13-1.el9.x86_64 83/153 2026-04-17T12:18:43.733 INFO:teuthology.orchestra.run.vm08.stdout: Installing : socat-1.7.4.1-8.el9.x86_64 49/153 2026-04-17T12:18:43.740 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-toml-0.10.2-6.el9.0.1.noarch 50/153 2026-04-17T12:18:43.752 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-functools-3.5.0-2.el9.noarch 51/153 2026-04-17T12:18:43.758 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-tempora-5.0.0-2.el9.noarch 52/153 2026-04-17T12:18:43.806 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-portend-3.1.0-2.el9.noarch 53/153 2026-04-17T12:18:43.816 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-protobuf-3.14.0-17.el9_7.noarch 54/153 2026-04-17T12:18:43.826 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-markupsafe-1.1.1-12.el9.x86_64 55/153 2026-04-17T12:18:43.874 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jmespath-1.0.1-1.el9_7.noarch 56/153 2026-04-17T12:18:44.073 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.13-1.el9.x86_64 83/153 2026-04-17T12:18:44.073 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-04-17T12:18:44.073 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:44.086 INFO:teuthology.orchestra.run.vm06.stdout: Installing : mailcap-2.1.49-5.el9.0.2.noarch 84/153 2026-04-17T12:18:44.091 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 85/153 2026-04-17T12:18:44.132 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:44.132 INFO:teuthology.orchestra.run.vm06.stdout:Creating group 'libstoragemgmt' with GID 992. 2026-04-17T12:18:44.132 INFO:teuthology.orchestra.run.vm06.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 992 and GID 992. 2026-04-17T12:18:44.132 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:44.148 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:44.198 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:44.198 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-04-17T12:18:44.198 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:44.230 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 87/153 2026-04-17T12:18:44.244 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-devel-3.9.23-2.el9.x86_64 57/153 2026-04-17T12:18:44.276 INFO:teuthology.orchestra.run.vm06.stdout: Installing : fuse-2.9.9-17.el9.x86_64 88/153 2026-04-17T12:18:44.284 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-babel-2.9.1-2.el9.noarch 58/153 2026-04-17T12:18:44.285 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cryptsetup-2.7.2-4.el9.x86_64 89/153 2026-04-17T12:18:44.290 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jinja2-2.11.3-8.el9_5.noarch 59/153 2026-04-17T12:18:44.293 INFO:teuthology.orchestra.run.vm06.stdout: Installing : c-ares-1.19.1-2.el9_4.x86_64 90/153 2026-04-17T12:18:44.297 INFO:teuthology.orchestra.run.vm08.stdout: Installing : perl-Benchmark-1.23-481.1.el9_6.noarch 60/153 2026-04-17T12:18:44.299 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 91/153 2026-04-17T12:18:44.323 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 92/153 2026-04-17T12:18:44.334 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 93/153 2026-04-17T12:18:44.371 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-0.3.29-1.el9.x86_64 61/153 2026-04-17T12:18:44.375 INFO:teuthology.orchestra.run.vm08.stdout: Installing : openblas-openmp-0.3.29-1.el9.x86_64 62/153 2026-04-17T12:18:44.412 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 63/153 2026-04-17T12:18:44.413 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 94/153 2026-04-17T12:18:44.430 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 95/153 2026-04-17T12:18:44.455 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 96/153 2026-04-17T12:18:44.476 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 97/153 2026-04-17T12:18:44.491 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 98/153 2026-04-17T12:18:44.529 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 99/153 2026-04-17T12:18:44.550 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 100/153 2026-04-17T12:18:44.577 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 101/153 2026-04-17T12:18:44.587 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 102/153 2026-04-17T12:18:44.601 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 103/153 2026-04-17T12:18:44.687 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 104/153 2026-04-17T12:18:44.950 INFO:teuthology.orchestra.run.vm08.stdout: Installing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 64/153 2026-04-17T12:18:45.065 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-1:1.23.5-2.el9_7.x86_64 65/153 2026-04-17T12:18:45.206 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 105/153 2026-04-17T12:18:45.236 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 106/153 2026-04-17T12:18:45.243 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 107/153 2026-04-17T12:18:45.253 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 108/153 2026-04-17T12:18:45.262 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 109/153 2026-04-17T12:18:45.268 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 110/153 2026-04-17T12:18:45.311 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 111/153 2026-04-17T12:18:45.319 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-cherrypy-18.10.0-5.el9.noarch 112/153 2026-04-17T12:18:45.329 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 113/153 2026-04-17T12:18:45.335 INFO:teuthology.orchestra.run.vm06.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 114/153 2026-04-17T12:18:45.345 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 115/153 2026-04-17T12:18:45.729 INFO:teuthology.orchestra.run.vm06.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 116/153 2026-04-17T12:18:45.748 INFO:teuthology.orchestra.run.vm06.stdout: Installing : librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 117/153 2026-04-17T12:18:45.778 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 117/153 2026-04-17T12:18:45.785 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 118/153 2026-04-17T12:18:45.972 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 66/153 2026-04-17T12:18:45.995 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-scipy-1.9.3-2.el9.x86_64 67/153 2026-04-17T12:18:46.011 INFO:teuthology.orchestra.run.vm08.stdout: Installing : boost-program-options-1.75.0-13.el9_7.x86_64 68/153 2026-04-17T12:18:46.015 INFO:teuthology.orchestra.run.vm08.stdout: Installing : smartmontools-1:7.2-9.el9.x86_64 69/153 2026-04-17T12:18:46.032 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: smartmontools-1:7.2-9.el9.x86_64 69/153 2026-04-17T12:18:46.032 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartd.service → /usr/lib/systemd/system/smartd.service. 2026-04-17T12:18:46.032 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:46.053 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 70/153 2026-04-17T12:18:46.064 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-packaging-20.9-5.el9.noarch 71/153 2026-04-17T12:18:46.086 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-ply-3.11-14.el9.0.1.noarch 72/153 2026-04-17T12:18:46.115 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pycparser-2.20-6.el9.noarch 73/153 2026-04-17T12:18:46.229 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cffi-1.14.5-5.el9.x86_64 74/153 2026-04-17T12:18:46.245 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cryptography-36.0.1-5.el9_6.x86_64 75/153 2026-04-17T12:18:46.279 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-pyOpenSSL-21.0.0-1.el9.noarch 76/153 2026-04-17T12:18:46.288 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cheroot-10.0.1-5.el9.noarch 77/153 2026-04-17T12:18:46.295 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-bcrypt-3.2.2-1.el9.x86_64 78/153 2026-04-17T12:18:46.298 INFO:teuthology.orchestra.run.vm08.stdout: Installing : pciutils-3.7.0-7.el9.x86_64 79/153 2026-04-17T12:18:46.335 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-24.09.0-1.el9.x86_64 80/153 2026-04-17T12:18:46.340 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-24.09.0-1.el9.x86_64 80/153 2026-04-17T12:18:46.342 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatlib-service-24.09.0-1.el9.x86_64 81/153 2026-04-17T12:18:46.361 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: qatlib-service-24.09.0-1.el9.x86_64 81/153 2026-04-17T12:18:46.511 INFO:teuthology.orchestra.run.vm08.stdout: Installing : qatzip-libs-1.3.1-1.el9.x86_64 82/153 2026-04-17T12:18:46.516 INFO:teuthology.orchestra.run.vm08.stdout: Installing : nvme-cli-2.13-1.el9.x86_64 83/153 2026-04-17T12:18:46.885 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: nvme-cli-2.13-1.el9.x86_64 83/153 2026-04-17T12:18:46.885 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-04-17T12:18:46.885 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:46.892 INFO:teuthology.orchestra.run.vm08.stdout: Installing : mailcap-2.1.49-5.el9.0.2.noarch 84/153 2026-04-17T12:18:46.896 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libconfig-1.7.2-9.el9.x86_64 85/153 2026-04-17T12:18:46.934 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:46.935 INFO:teuthology.orchestra.run.vm08.stdout:Creating group 'libstoragemgmt' with GID 992. 2026-04-17T12:18:46.935 INFO:teuthology.orchestra.run.vm08.stdout:Creating user 'libstoragemgmt' (daemon account for libstoragemgmt) with UID 992 and GID 992. 2026-04-17T12:18:46.935 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:46.949 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:46.996 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 86/153 2026-04-17T12:18:46.996 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/libstoragemgmt.service → /usr/lib/systemd/system/libstoragemgmt.service. 2026-04-17T12:18:46.996 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:47.026 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 87/153 2026-04-17T12:18:47.063 INFO:teuthology.orchestra.run.vm08.stdout: Installing : fuse-2.9.9-17.el9.x86_64 88/153 2026-04-17T12:18:47.071 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cryptsetup-2.7.2-4.el9.x86_64 89/153 2026-04-17T12:18:47.081 INFO:teuthology.orchestra.run.vm08.stdout: Installing : c-ares-1.19.1-2.el9_4.x86_64 90/153 2026-04-17T12:18:47.089 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-zc-lockfile-2.0-10.el9.noarch 91/153 2026-04-17T12:18:47.114 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-xmltodict-0.12.0-15.el9.noarch 92/153 2026-04-17T12:18:47.123 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-websocket-client-1.2.3-2.el9.noarch 93/153 2026-04-17T12:18:47.191 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-typing-extensions-4.15.0-1.el9.noarch 94/153 2026-04-17T12:18:47.205 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-asyncssh-2.13.2-5.el9.noarch 95/153 2026-04-17T12:18:47.222 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-repoze-lru-0.7-16.el9.noarch 96/153 2026-04-17T12:18:47.237 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-routes-2.5.1-5.el9.noarch 97/153 2026-04-17T12:18:47.246 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-natsort-7.1.1-5.el9.noarch 98/153 2026-04-17T12:18:47.281 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-msgpack-1.0.3-2.el9.x86_64 99/153 2026-04-17T12:18:47.302 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-influxdb-5.3.1-1.el9.noarch 100/153 2026-04-17T12:18:47.328 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:47.334 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:47.335 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-isodate-0.6.1-3.el9.noarch 101/153 2026-04-17T12:18:47.344 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-saml-1.16.0-1.el9.noarch 102/153 2026-04-17T12:18:47.356 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-certifi-2023.05.07-4.el9.noarch 103/153 2026-04-17T12:18:47.422 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cachetools-4.2.4-1.el9.noarch 104/153 2026-04-17T12:18:47.750 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:47.758 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 120/153 2026-04-17T12:18:47.811 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 120/153 2026-04-17T12:18:47.811 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-04-17T12:18:47.811 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-04-17T12:18:47.811 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:47.827 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 121/153 2026-04-17T12:18:47.906 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-google-auth-1:2.45.0-1.el9.noarch 105/153 2026-04-17T12:18:47.930 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-kubernetes-1:26.1.0-3.el9.noarch 106/153 2026-04-17T12:18:47.940 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-backports-tarfile-1.2.0-1.el9.noarch 107/153 2026-04-17T12:18:47.950 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-context-6.0.1-3.el9.noarch 108/153 2026-04-17T12:18:47.962 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-autocommand-2.2.2-8.el9.noarch 109/153 2026-04-17T12:18:47.968 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-text-4.0.0-2.el9.noarch 110/153 2026-04-17T12:18:48.005 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-jaraco-collections-3.0.0-8.el9.noarch 111/153 2026-04-17T12:18:48.012 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-cherrypy-18.10.0-5.el9.noarch 112/153 2026-04-17T12:18:48.020 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libunwind-1.6.2-1.el9.x86_64 113/153 2026-04-17T12:18:48.024 INFO:teuthology.orchestra.run.vm08.stdout: Installing : gperftools-libs-2.9.1-3.el9.x86_64 114/153 2026-04-17T12:18:48.031 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libarrow-doc-9.0.0-15.el9.noarch 115/153 2026-04-17T12:18:48.396 INFO:teuthology.orchestra.run.vm08.stdout: Installing : parquet-libs-9.0.0-15.el9.x86_64 116/153 2026-04-17T12:18:48.463 INFO:teuthology.orchestra.run.vm08.stdout: Installing : librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 117/153 2026-04-17T12:18:48.487 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 117/153 2026-04-17T12:18:48.490 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 118/153 2026-04-17T12:18:49.906 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:49.911 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:50.315 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 119/153 2026-04-17T12:18:50.323 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 120/153 2026-04-17T12:18:50.374 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 120/153 2026-04-17T12:18:50.374 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /usr/lib/systemd/system/ceph.target. 2026-04-17T12:18:50.375 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /usr/lib/systemd/system/ceph-crash.service. 2026-04-17T12:18:50.375 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:50.379 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 121/153 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 121/153 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /sys 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /proc 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /mnt 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /var/tmp 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /home 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /root 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout:skipping the directory /tmp 2026-04-17T12:18:50.398 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:50.518 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 122/153 2026-04-17T12:18:50.543 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 122/153 2026-04-17T12:18:50.543 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:50.543 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-17T12:18:50.543 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:51.450 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 123/153 2026-04-17T12:18:51.479 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 123/153 2026-04-17T12:18:51.480 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:51.480 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-17T12:18:51.480 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-17T12:18:51.480 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-17T12:18:51.480 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:51.611 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 124/153 2026-04-17T12:18:51.642 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 124/153 2026-04-17T12:18:51.642 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:51.642 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-17T12:18:51.642 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-17T12:18:51.642 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-17T12:18:51.642 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:51.952 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 125/153 2026-04-17T12:18:51.978 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 125/153 2026-04-17T12:18:51.978 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:51.978 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-17T12:18:51.978 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-17T12:18:51.978 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-17T12:18:51.979 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:52.029 INFO:teuthology.orchestra.run.vm02.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 126/153 2026-04-17T12:18:52.095 INFO:teuthology.orchestra.run.vm02.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 127/153 2026-04-17T12:18:52.116 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 128/153 2026-04-17T12:18:52.122 INFO:teuthology.orchestra.run.vm02.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 129/153 2026-04-17T12:18:52.221 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 130/153 2026-04-17T12:18:52.226 INFO:teuthology.orchestra.run.vm02.stdout: Installing : cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 130/153 2026-04-17T12:18:52.238 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 131/153 2026-04-17T12:18:52.281 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 132/153 2026-04-17T12:18:52.285 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 133/153 2026-04-17T12:18:54.365 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 133/153 2026-04-17T12:18:54.463 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 134/153 2026-04-17T12:18:55.091 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 134/153 2026-04-17T12:18:55.094 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 135/153 2026-04-17T12:18:55.115 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 135/153 2026-04-17T12:18:55.117 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 136/153 2026-04-17T12:18:55.199 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 136/153 2026-04-17T12:18:55.256 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 137/153 2026-04-17T12:18:55.261 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 138/153 2026-04-17T12:18:55.294 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 138/153 2026-04-17T12:18:55.294 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:55.294 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-17T12:18:55.294 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-17T12:18:55.294 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-17T12:18:55.294 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:55.311 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 139/153 2026-04-17T12:18:55.330 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 139/153 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 121/153 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /sys 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /proc 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /mnt 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /var/tmp 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /home 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /root 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout:skipping the directory /tmp 2026-04-17T12:18:55.531 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:55.637 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 122/153 2026-04-17T12:18:55.660 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 122/153 2026-04-17T12:18:55.660 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:55.660 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-17T12:18:55.660 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:56.646 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 123/153 2026-04-17T12:18:56.686 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 123/153 2026-04-17T12:18:56.687 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:56.687 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-17T12:18:56.687 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-17T12:18:56.687 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-17T12:18:56.687 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:56.743 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 140/153 2026-04-17T12:18:56.749 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 141/153 2026-04-17T12:18:56.786 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 141/153 2026-04-17T12:18:56.786 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:56.786 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-17T12:18:56.786 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-17T12:18:56.786 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-17T12:18:56.786 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:56.800 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 142/153 2026-04-17T12:18:56.832 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 124/153 2026-04-17T12:18:56.833 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 142/153 2026-04-17T12:18:56.833 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:56.833 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-17T12:18:56.833 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:56.869 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 124/153 2026-04-17T12:18:56.869 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:56.869 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-17T12:18:56.870 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-17T12:18:56.870 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-17T12:18:56.870 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:57.079 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 143/153 2026-04-17T12:18:57.107 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 143/153 2026-04-17T12:18:57.107 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:57.107 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-17T12:18:57.107 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-17T12:18:57.107 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-17T12:18:57.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:18:57.349 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 125/153 2026-04-17T12:18:57.380 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 125/153 2026-04-17T12:18:57.380 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:57.380 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-17T12:18:57.380 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-17T12:18:57.380 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-17T12:18:57.380 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:18:57.434 INFO:teuthology.orchestra.run.vm06.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 126/153 2026-04-17T12:18:57.512 INFO:teuthology.orchestra.run.vm06.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 127/153 2026-04-17T12:18:57.667 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 128/153 2026-04-17T12:18:57.672 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 129/153 2026-04-17T12:18:57.764 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 130/153 2026-04-17T12:18:57.770 INFO:teuthology.orchestra.run.vm06.stdout: Installing : cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 130/153 2026-04-17T12:18:57.781 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 131/153 2026-04-17T12:18:57.823 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 132/153 2026-04-17T12:18:57.827 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 133/153 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 121/153 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /sys 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /proc 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /mnt 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /var/tmp 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /home 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /root 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout:skipping the directory /tmp 2026-04-17T12:18:58.223 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:58.324 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 122/153 2026-04-17T12:18:58.348 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 122/153 2026-04-17T12:18:58.348 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:58.348 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-17T12:18:58.348 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:59.280 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 123/153 2026-04-17T12:18:59.320 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 123/153 2026-04-17T12:18:59.320 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:59.320 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-17T12:18:59.320 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-17T12:18:59.320 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /usr/lib/systemd/system/ceph-osd.target. 2026-04-17T12:18:59.320 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:59.471 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 124/153 2026-04-17T12:18:59.497 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 124/153 2026-04-17T12:18:59.498 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:59.498 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-17T12:18:59.498 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-17T12:18:59.498 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /usr/lib/systemd/system/ceph-mds.target. 2026-04-17T12:18:59.498 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:59.793 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 125/153 2026-04-17T12:18:59.825 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 125/153 2026-04-17T12:18:59.825 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:18:59.825 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-17T12:18:59.825 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-17T12:18:59.825 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /usr/lib/systemd/system/ceph-mon.target. 2026-04-17T12:18:59.826 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:18:59.867 INFO:teuthology.orchestra.run.vm08.stdout: Installing : grpc-data-1.46.7-10.el9.noarch 126/153 2026-04-17T12:18:59.956 INFO:teuthology.orchestra.run.vm08.stdout: Installing : abseil-cpp-20211102.0-4.el9.x86_64 127/153 2026-04-17T12:18:59.971 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 133/153 2026-04-17T12:18:59.980 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-1.46.7-10.el9.x86_64 128/153 2026-04-17T12:18:59.987 INFO:teuthology.orchestra.run.vm08.stdout: Installing : python3-grpcio-tools-1.46.7-10.el9.x86_64 129/153 2026-04-17T12:18:59.987 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 134/153 2026-04-17T12:19:00.069 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 130/153 2026-04-17T12:19:00.074 INFO:teuthology.orchestra.run.vm08.stdout: Installing : cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 130/153 2026-04-17T12:19:00.086 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 131/153 2026-04-17T12:19:00.122 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 132/153 2026-04-17T12:19:00.126 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 133/153 2026-04-17T12:19:00.591 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 134/153 2026-04-17T12:19:00.594 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 135/153 2026-04-17T12:19:00.609 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 135/153 2026-04-17T12:19:00.612 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 136/153 2026-04-17T12:19:00.682 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 136/153 2026-04-17T12:19:00.740 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 137/153 2026-04-17T12:19:00.743 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 138/153 2026-04-17T12:19:00.770 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 138/153 2026-04-17T12:19:00.770 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:00.770 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-17T12:19:00.770 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-17T12:19:00.770 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-17T12:19:00.770 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:00.786 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 139/153 2026-04-17T12:19:00.800 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 139/153 2026-04-17T12:19:02.088 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 144/153 2026-04-17T12:19:02.129 INFO:teuthology.orchestra.run.vm02.stdout: Installing : ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 145/153 2026-04-17T12:19:02.138 INFO:teuthology.orchestra.run.vm02.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 146/153 2026-04-17T12:19:02.141 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 140/153 2026-04-17T12:19:02.145 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 141/153 2026-04-17T12:19:02.146 INFO:teuthology.orchestra.run.vm02.stdout: Installing : libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.cly 147/153 2026-04-17T12:19:02.149 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 133/153 2026-04-17T12:19:02.160 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 148/153 2026-04-17T12:19:02.166 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 134/153 2026-04-17T12:19:02.172 INFO:teuthology.orchestra.run.vm02.stdout: Installing : rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 149/153 2026-04-17T12:19:02.179 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 141/153 2026-04-17T12:19:02.179 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:02.179 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-17T12:19:02.179 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-17T12:19:02.179 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-17T12:19:02.179 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:02.194 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 142/153 2026-04-17T12:19:02.198 INFO:teuthology.orchestra.run.vm02.stdout: Installing : bzip2-1.0.8-10.el9_5.x86_64 150/153 2026-04-17T12:19:02.203 INFO:teuthology.orchestra.run.vm02.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 151/153 2026-04-17T12:19:02.203 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 152/153 2026-04-17T12:19:02.223 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 152/153 2026-04-17T12:19:02.223 INFO:teuthology.orchestra.run.vm02.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:02.227 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 142/153 2026-04-17T12:19:02.227 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:02.227 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-17T12:19:02.227 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:02.399 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 143/153 2026-04-17T12:19:02.434 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 143/153 2026-04-17T12:19:02.434 INFO:teuthology.orchestra.run.vm06.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:02.434 INFO:teuthology.orchestra.run.vm06.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-17T12:19:02.434 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-17T12:19:02.434 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-17T12:19:02.434 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:02.856 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 134/153 2026-04-17T12:19:02.877 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 135/153 2026-04-17T12:19:02.898 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 135/153 2026-04-17T12:19:02.900 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 136/153 2026-04-17T12:19:02.975 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 136/153 2026-04-17T12:19:03.030 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 137/153 2026-04-17T12:19:03.034 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 138/153 2026-04-17T12:19:03.063 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 138/153 2026-04-17T12:19:03.063 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:03.063 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-17T12:19:03.063 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-17T12:19:03.063 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /usr/lib/systemd/system/ceph-mgr.target. 2026-04-17T12:19:03.063 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:03.079 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 139/153 2026-04-17T12:19:03.097 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 139/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 2/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 3/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 5/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 6/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 7/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 8/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 9/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 10/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 11/153 2026-04-17T12:19:03.924 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 12/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.cl 13/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.cly 14/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 15/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 16/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 17/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clys 18/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 19/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 20/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.e 21/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 22/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clys 23/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso 24/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 25/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 26/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 27/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 28/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 29/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 30/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 31/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 32/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 33/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 34/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 35/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 36/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 37/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 38/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 39/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 40/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 41/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 42/153 2026-04-17T12:19:03.925 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 43/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 44/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 45/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 46/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 47/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 48/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 49/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 50/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 51/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 52/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 53/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 54/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cheroot-10.0.1-5.el9.noarch 55/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cherrypy-18.10.0-5.el9.noarch 56/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 57/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 58/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 59/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 60/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 61/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 62/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 63/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 64/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 65/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 66/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 67/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 68/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 69/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 70/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 71/153 2026-04-17T12:19:03.929 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 72/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 73/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 74/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 75/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 76/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 77/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 78/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 79/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 80/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 81/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 82/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 83/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : re2-1:20211101-20.el9.x86_64 84/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 85/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 86/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : bzip2-1.0.8-10.el9_5.x86_64 87/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : c-ares-1.19.1-2.el9_4.x86_64 88/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : cryptsetup-2.7.2-4.el9.x86_64 89/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 90/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 91/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 92/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libgfortran-11.5.0-11.el9.x86_64 93/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libquadmath-11.5.0-11.el9.x86_64 94/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 95/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : nvme-cli-2.13-1.el9.x86_64 96/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 97/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 98/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-cryptography-36.0.1-5.el9_6.x86_64 99/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-ply-3.11-14.el9.0.1.noarch 100/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 101/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.0.1.noarch 102/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-2.25.1-10.el9_6.noarch 103/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-urllib3-1.26.5-6.el9_7.1.noarch 104/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : smartmontools-1:7.2-9.el9.x86_64 105/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 106/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 107/153 2026-04-17T12:19:03.930 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : boost-program-options-1.75.0-13.el9_7.x86_64 108/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-3.0.4-8.el9.0.1.x86_64 109/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 110/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 111/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 112/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 113/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 114/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 115/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 116/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : libxslt-1.1.34-14.el9_7.1.x86_64 117/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 118/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 119/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 120/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 121/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : perl-Benchmark-1.23-481.1.el9_6.noarch 122/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 123/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-3.14.0-17.el9_7.x86_64 124/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 125/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-devel-3.9.23-2.el9.x86_64 126/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jinja2-2.11.3-8.el9_5.noarch 127/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-jmespath-1.0.1-1.el9_7.noarch 128/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 129/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 130/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 131/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-1:1.23.5-2.el9_7.x86_64 132/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 133/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 134/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-protobuf-3.14.0-17.el9_7.noarch 135/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-0.4.8-7.el9_7.noarch 136/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 137/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 138/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 139/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : python3-toml-0.10.2-6.el9.0.1.noarch 140/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-24.09.0-1.el9.x86_64 141/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatlib-service-24.09.0-1.el9.x86_64 142/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 143/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 144/153 2026-04-17T12:19:03.931 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 145/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 146/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 147/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 148/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : protobuf-compiler-3.14.0-17.el9_7.x86_64 149/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 150/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 151/153 2026-04-17T12:19:03.932 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 152/153 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout:Upgraded: 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout:Installed: 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: boost-program-options-1.75.0-13.el9_7.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: bzip2-1.0.8-10.el9_5.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: c-ares-1.19.1-2.el9_4.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.041 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup-2.7.2-4.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: fuse-2.9.9-17.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data-1.46.7-10.el9.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-9.0.0-15.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libconfig-1.7.2-9.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran-11.5.0-11.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libnbd-1.20.3-4.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: liboath-2.6.12-1.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath-11.5.0-11.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libunwind-1.6.2-1.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: libxslt-1.1.34-14.el9_7.1.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-17T12:19:04.042 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli-2.13-1.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: openblas-0.3.29-1.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: pciutils-3.7.0-7.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: perl-Benchmark-1.23-481.1.el9_6.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-3.14.0-17.el9_7.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler-3.14.0-17.el9_7.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel-2.9.1-2.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot-10.0.1-5.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy-18.10.0-5.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography-36.0.1-5.el9_6.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel-3.9.23-2.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2-2.11.3-8.el9_5.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-jmespath-1.0.1-1.el9_7.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-1:1.23.5-2.el9_7.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging-20.9-5.el9.noarch 2026-04-17T12:19:04.043 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply-3.11-14.el9.0.1.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend-3.1.0-2.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf-3.14.0-17.el9_7.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-0.4.8-7.el9_7.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules-0.4.8-7.el9_7.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser-2.20-6.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing-2.4.7-9.el9.0.1.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-2.25.1-10.el9_6.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes-2.5.1-5.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa-4.9-2.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml-1.16.0-1.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml-0.10.2-6.el9.0.1.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3-1.26.5-6.el9_7.1.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-24.09.0-1.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service-24.09.0-1.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: re2-1:20211101-20.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: s3cmd-2.4.0-1.el9.noarch 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools-1:7.2-9.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: thrift-0.15.0-4.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:04.044 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-04-17T12:19:04.132 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:04.424 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 140/153 2026-04-17T12:19:04.428 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 141/153 2026-04-17T12:19:04.458 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 141/153 2026-04-17T12:19:04.458 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:04.458 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-17T12:19:04.458 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-17T12:19:04.458 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /usr/lib/systemd/system/ceph-radosgw.target. 2026-04-17T12:19:04.458 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:04.473 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 142/153 2026-04-17T12:19:04.498 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 142/153 2026-04-17T12:19:04.498 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:04.498 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-17T12:19:04.498 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:04.658 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 143/153 2026-04-17T12:19:04.683 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 143/153 2026-04-17T12:19:04.683 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:19:04.683 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-17T12:19:04.683 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-17T12:19:04.683 INFO:teuthology.orchestra.run.vm08.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target → /usr/lib/systemd/system/ceph-rbd-mirror.target. 2026-04-17T12:19:04.684 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:07.350 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 144/153 2026-04-17T12:19:07.387 INFO:teuthology.orchestra.run.vm06.stdout: Installing : ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 145/153 2026-04-17T12:19:07.396 INFO:teuthology.orchestra.run.vm06.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 146/153 2026-04-17T12:19:07.405 INFO:teuthology.orchestra.run.vm06.stdout: Installing : libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.cly 147/153 2026-04-17T12:19:07.420 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 148/153 2026-04-17T12:19:07.433 INFO:teuthology.orchestra.run.vm06.stdout: Installing : rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 149/153 2026-04-17T12:19:07.456 INFO:teuthology.orchestra.run.vm06.stdout: Installing : bzip2-1.0.8-10.el9_5.x86_64 150/153 2026-04-17T12:19:07.461 INFO:teuthology.orchestra.run.vm06.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 151/153 2026-04-17T12:19:07.461 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 152/153 2026-04-17T12:19:07.485 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 152/153 2026-04-17T12:19:07.485 INFO:teuthology.orchestra.run.vm06.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 2/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 3/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 5/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 6/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 7/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 8/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 9/153 2026-04-17T12:19:09.368 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 10/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 11/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 12/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.cl 13/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.cly 14/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 15/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 16/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 17/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clys 18/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 19/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 20/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.e 21/153 2026-04-17T12:19:09.369 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 22/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clys 23/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso 24/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 25/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 26/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 27/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 28/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 29/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 30/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 31/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 32/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 33/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 34/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 35/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 36/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 37/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 38/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 39/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 40/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 41/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 42/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 43/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 44/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 45/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 46/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 47/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 48/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 49/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 50/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 51/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 52/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 53/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 54/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cheroot-10.0.1-5.el9.noarch 55/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cherrypy-18.10.0-5.el9.noarch 56/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 57/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 58/153 2026-04-17T12:19:09.370 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 59/153 2026-04-17T12:19:09.372 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 60/153 2026-04-17T12:19:09.372 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 61/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 62/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 63/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 64/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 65/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 66/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 67/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 68/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 69/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 70/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 71/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 72/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 73/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 74/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 75/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 76/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 77/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 78/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 79/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 80/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 81/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 82/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 83/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : re2-1:20211101-20.el9.x86_64 84/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 85/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 86/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : bzip2-1.0.8-10.el9_5.x86_64 87/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : c-ares-1.19.1-2.el9_4.x86_64 88/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : cryptsetup-2.7.2-4.el9.x86_64 89/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 90/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 91/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 92/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libgfortran-11.5.0-11.el9.x86_64 93/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libquadmath-11.5.0-11.el9.x86_64 94/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 95/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.13-1.el9.x86_64 96/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 97/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 98/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-cryptography-36.0.1-5.el9_6.x86_64 99/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-ply-3.11-14.el9.0.1.noarch 100/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 101/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.0.1.noarch 102/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-2.25.1-10.el9_6.noarch 103/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-6.el9_7.1.noarch 104/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : smartmontools-1:7.2-9.el9.x86_64 105/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : unzip-6.0-59.el9.x86_64 106/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : zip-3.0-35.el9.x86_64 107/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : boost-program-options-1.75.0-13.el9_7.x86_64 108/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-3.0.4-8.el9.0.1.x86_64 109/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 110/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 111/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 112/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 113/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 114/153 2026-04-17T12:19:09.373 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 115/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 116/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : libxslt-1.1.34-14.el9_7.1.x86_64 117/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 118/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-5.4.4-4.el9.x86_64 119/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 120/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 121/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : perl-Benchmark-1.23-481.1.el9_6.noarch 122/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 123/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-3.14.0-17.el9_7.x86_64 124/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 125/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-devel-3.9.23-2.el9.x86_64 126/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jinja2-2.11.3-8.el9_5.noarch 127/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9_7.noarch 128/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 129/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 130/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 131/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-1:1.23.5-2.el9_7.x86_64 132/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 133/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 134/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-protobuf-3.14.0-17.el9_7.noarch 135/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-0.4.8-7.el9_7.noarch 136/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 137/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 138/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 139/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-toml-0.10.2-6.el9.0.1.noarch 140/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-24.09.0-1.el9.x86_64 141/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatlib-service-24.09.0-1.el9.x86_64 142/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 143/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 144/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 145/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 146/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 147/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 148/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : protobuf-compiler-3.14.0-17.el9_7.x86_64 149/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 150/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 151/153 2026-04-17T12:19:09.374 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 152/153 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout:Upgraded: 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: boost-program-options-1.75.0-13.el9_7.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: bzip2-1.0.8-10.el9_5.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: c-ares-1.19.1-2.el9_4.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.494 INFO:teuthology.orchestra.run.vm06.stdout: cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: cryptsetup-2.7.2-4.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: fuse-2.9.9-17.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: grpc-data-1.46.7-10.el9.noarch 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-9.0.0-15.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libconfig-1.7.2-9.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libgfortran-11.5.0-11.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libnbd-1.20.3-4.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: liboath-2.6.12-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libquadmath-11.5.0-11.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libunwind-1.6.2-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: libxslt-1.1.34-14.el9_7.1.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: lua-5.4.4-4.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.13-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: openblas-0.3.29-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: pciutils-3.7.0-7.el9.x86_64 2026-04-17T12:19:09.495 INFO:teuthology.orchestra.run.vm06.stdout: perl-Benchmark-1.23-481.1.el9_6.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-3.14.0-17.el9_7.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: protobuf-compiler-3.14.0-17.el9_7.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-babel-2.9.1-2.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-cheroot-10.0.1-5.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-cherrypy-18.10.0-5.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-cryptography-36.0.1-5.el9_6.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-devel-3.9.23-2.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jinja2-2.11.3-8.el9_5.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9_7.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-1:1.23.5-2.el9_7.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-packaging-20.9-5.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-ply-3.11-14.el9.0.1.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-portend-3.1.0-2.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-protobuf-3.14.0-17.el9_7.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-0.4.8-7.el9_7.noarch 2026-04-17T12:19:09.496 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyasn1-modules-0.4.8-7.el9_7.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-pycparser-2.20-6.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.0.1.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-2.25.1-10.el9_6.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-routes-2.5.1-5.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-rsa-4.9-2.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-saml-1.16.0-1.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-toml-0.10.2-6.el9.0.1.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-6.el9_7.1.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-24.09.0-1.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: qatlib-service-24.09.0-1.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: re2-1:20211101-20.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: s3cmd-2.4.0-1.el9.noarch 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: smartmontools-1:7.2-9.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: thrift-0.15.0-4.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: unzip-6.0-59.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-04-17T12:19:09.497 INFO:teuthology.orchestra.run.vm06.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-17T12:19:09.498 INFO:teuthology.orchestra.run.vm06.stdout: zip-3.0-35.el9.x86_64 2026-04-17T12:19:09.498 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:09.498 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-04-17T12:19:09.583 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 144/153 2026-04-17T12:19:09.605 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:09.624 INFO:teuthology.orchestra.run.vm08.stdout: Installing : ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 145/153 2026-04-17T12:19:09.635 INFO:teuthology.orchestra.run.vm08.stdout: Installing : perl-Test-Harness-1:3.42-461.el9.noarch 146/153 2026-04-17T12:19:09.644 INFO:teuthology.orchestra.run.vm08.stdout: Installing : libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.cly 147/153 2026-04-17T12:19:09.661 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 148/153 2026-04-17T12:19:09.672 INFO:teuthology.orchestra.run.vm08.stdout: Installing : rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 149/153 2026-04-17T12:19:09.696 INFO:teuthology.orchestra.run.vm08.stdout: Installing : bzip2-1.0.8-10.el9_5.x86_64 150/153 2026-04-17T12:19:09.702 INFO:teuthology.orchestra.run.vm08.stdout: Installing : s3cmd-2.4.0-1.el9.noarch 151/153 2026-04-17T12:19:09.702 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librbd1-2:16.2.4-5.el9.x86_64 152/153 2026-04-17T12:19:09.723 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librbd1-2:16.2.4-5.el9.x86_64 152/153 2026-04-17T12:19:09.723 INFO:teuthology.orchestra.run.vm08.stdout: Cleanup : librados2-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: librados2-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 2/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 3/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 4/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 5/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 6/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 7/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 8/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 9/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 10/153 2026-04-17T12:19:11.423 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 11/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 12/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.cl 13/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.cly 14/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.cl 15/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 16/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso 17/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clys 18/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.cl 19/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 20/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.e 21/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 22/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clys 23/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso 24/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 25/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 26/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 27/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 28/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 29/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 30/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 31/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 32/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 33/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 34/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 35/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 36/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 37/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 38/153 2026-04-17T12:19:11.424 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarc 39/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : abseil-cpp-20211102.0-4.el9.x86_64 40/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : gperftools-libs-2.9.1-3.el9.x86_64 41/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : grpc-data-1.46.7-10.el9.noarch 42/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-9.0.0-15.el9.x86_64 43/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libarrow-doc-9.0.0-15.el9.noarch 44/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : liboath-2.6.12-1.el9.x86_64 45/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libunwind-1.6.2-1.el9.x86_64 46/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 47/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : parquet-libs-9.0.0-15.el9.x86_64 48/153 2026-04-17T12:19:11.427 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-asyncssh-2.13.2-5.el9.noarch 49/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-autocommand-2.2.2-8.el9.noarch 50/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-backports-tarfile-1.2.0-1.el9.noarch 51/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-bcrypt-3.2.2-1.el9.x86_64 52/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cachetools-4.2.4-1.el9.noarch 53/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-certifi-2023.05.07-4.el9.noarch 54/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cheroot-10.0.1-5.el9.noarch 55/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cherrypy-18.10.0-5.el9.noarch 56/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-google-auth-1:2.45.0-1.el9.noarch 57/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-1.46.7-10.el9.x86_64 58/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-grpcio-tools-1.46.7-10.el9.x86_64 59/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-influxdb-5.3.1-1.el9.noarch 60/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-isodate-0.6.1-3.el9.noarch 61/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-8.2.1-3.el9.noarch 62/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-classes-3.2.1-5.el9.noarch 63/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-collections-3.0.0-8.el9.noarch 64/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-context-6.0.1-3.el9.noarch 65/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-functools-3.5.0-2.el9.noarch 66/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jaraco-text-4.0.0-2.el9.noarch 67/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-kubernetes-1:26.1.0-3.el9.noarch 68/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-more-itertools-8.12.0-2.el9.noarch 69/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-msgpack-1.0.3-2.el9.x86_64 70/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-natsort-7.1.1-5.el9.noarch 71/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-portend-3.1.0-2.el9.noarch 72/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyOpenSSL-21.0.0-1.el9.noarch 73/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-repoze-lru-0.7-16.el9.noarch 74/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-routes-2.5.1-5.el9.noarch 75/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-rsa-4.9-2.el9.noarch 76/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-saml-1.16.0-1.el9.noarch 77/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-tempora-5.0.0-2.el9.noarch 78/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-typing-extensions-4.15.0-1.el9.noarch 79/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-websocket-client-1.2.3-2.el9.noarch 80/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmlsec-1.3.13-1.el9.x86_64 81/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-xmltodict-0.12.0-15.el9.noarch 82/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-zc-lockfile-2.0-10.el9.noarch 83/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : re2-1:20211101-20.el9.x86_64 84/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : s3cmd-2.4.0-1.el9.noarch 85/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : thrift-0.15.0-4.el9.x86_64 86/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : bzip2-1.0.8-10.el9_5.x86_64 87/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : c-ares-1.19.1-2.el9_4.x86_64 88/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : cryptsetup-2.7.2-4.el9.x86_64 89/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : fuse-2.9.9-17.el9.x86_64 90/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ledmon-libs-1.1.0-3.el9.x86_64 91/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libconfig-1.7.2-9.el9.x86_64 92/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libgfortran-11.5.0-11.el9.x86_64 93/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libquadmath-11.5.0-11.el9.x86_64 94/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 95/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : nvme-cli-2.13-1.el9.x86_64 96/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : pciutils-3.7.0-7.el9.x86_64 97/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cffi-1.14.5-5.el9.x86_64 98/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-cryptography-36.0.1-5.el9_6.x86_64 99/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-ply-3.11-14.el9.0.1.noarch 100/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pycparser-2.20-6.el9.noarch 101/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.0.1.noarch 102/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-2.25.1-10.el9_6.noarch 103/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-urllib3-1.26.5-6.el9_7.1.noarch 104/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : smartmontools-1:7.2-9.el9.x86_64 105/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 106/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 107/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : boost-program-options-1.75.0-13.el9_7.x86_64 108/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-3.0.4-8.el9.0.1.x86_64 109/153 2026-04-17T12:19:11.428 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 110/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 111/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libnbd-1.20.3-4.el9.x86_64 112/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libpmemobj-1.12.1-1.el9.x86_64 113/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librabbitmq-0.11.0-7.el9.x86_64 114/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librdkafka-1.6.1-102.el9.x86_64 115/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libstoragemgmt-1.10.1-1.el9.x86_64 116/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : libxslt-1.1.34-14.el9_7.1.x86_64 117/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lttng-ust-2.12.0-6.el9.x86_64 118/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 119/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-0.3.29-1.el9.x86_64 120/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : openblas-openmp-0.3.29-1.el9.x86_64 121/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : perl-Benchmark-1.23-481.1.el9_6.noarch 122/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : perl-Test-Harness-1:3.42-461.el9.noarch 123/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-3.14.0-17.el9_7.x86_64 124/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-babel-2.9.1-2.el9.noarch 125/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-devel-3.9.23-2.el9.x86_64 126/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jinja2-2.11.3-8.el9_5.noarch 127/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-jmespath-1.0.1-1.el9_7.noarch 128/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-libstoragemgmt-1.10.1-1.el9.x86_64 129/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-lxml-4.6.5-3.el9.x86_64 130/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-markupsafe-1.1.1-12.el9.x86_64 131/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-1:1.23.5-2.el9_7.x86_64 132/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 133/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-packaging-20.9-5.el9.noarch 134/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-protobuf-3.14.0-17.el9_7.noarch 135/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-0.4.8-7.el9_7.noarch 136/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 137/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-requests-oauthlib-1.3.0-12.el9.noarch 138/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-scipy-1.9.3-2.el9.x86_64 139/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : python3-toml-0.10.2-6.el9.0.1.noarch 140/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-24.09.0-1.el9.x86_64 141/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatlib-service-24.09.0-1.el9.x86_64 142/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : qatzip-libs-1.3.1-1.el9.x86_64 143/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 144/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-1.2.29-13.el9.x86_64 145/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlsec1-openssl-1.2.29-13.el9.x86_64 146/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 147/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 148/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : protobuf-compiler-3.14.0-17.el9_7.x86_64 149/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 150/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librados2-2:16.2.4-5.el9.x86_64 151/153 2026-04-17T12:19:11.429 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_6 152/153 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : librbd1-2:16.2.4-5.el9.x86_64 153/153 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout:Upgraded: 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: librados2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: librbd1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout:Installed: 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp-20211102.0-4.el9.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: boost-program-options-1.75.0-13.el9_7.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: bzip2-1.0.8-10.el9_5.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: c-ares-1.19.1-2.el9_4.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: ceph-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.550 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: cephadm-2:20.2.0-21.gc03ba9ecf58.el9.clyso.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup-2.7.2-4.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: fuse-2.9.9-17.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs-2.9.1-3.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data-1.46.7-10.el9.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs-1.1.0-3.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-9.0.0-15.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libarrow-doc-9.0.0-15.el9.noarch 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-daemon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-devel-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs-proxy2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libcephfs2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libconfig-1.7.2-9.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran-11.5.0-11.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libnbd-1.20.3-4.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: liboath-2.6.12-1.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libpmemobj-1.12.1-1.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath-11.5.0-11.el9.x86_64 2026-04-17T12:19:11.551 INFO:teuthology.orchestra.run.vm08.stdout: librabbitmq-0.11.0-7.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: librados-devel-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: librdkafka-1.6.1-102.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: librgw2-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: libunwind-1.6.2-1.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: libxslt-1.1.34-14.el9_7.1.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: lttng-ust-2.12.0-6.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli-2.13-1.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: openblas-0.3.29-1.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp-0.3.29-1.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: parquet-libs-9.0.0-15.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: pciutils-3.7.0-7.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: perl-Benchmark-1.23-481.1.el9_6.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: perl-Test-Harness-1:3.42-461.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-3.14.0-17.el9_7.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler-3.14.0-17.el9_7.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh-2.13.2-5.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand-2.2.2-8.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel-2.9.1-2.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile-1.2.0-1.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt-3.2.2-1.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools-4.2.4-1.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-argparse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-cephfs-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi-2023.05.07-4.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi-1.14.5-5.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot-10.0.1-5.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy-18.10.0-5.el9.noarch 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography-36.0.1-5.el9_6.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel-3.9.23-2.el9.x86_64 2026-04-17T12:19:11.552 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth-1:2.45.0-1.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-1.46.7-10.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools-1.46.7-10.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb-5.3.1-1.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate-0.6.1-3.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-8.2.1-3.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes-3.2.1-5.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections-3.0.0-8.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context-6.0.1-3.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools-3.5.0-2.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text-4.0.0-2.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2-2.11.3-8.el9_5.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-jmespath-1.0.1-1.el9_7.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes-1:26.1.0-3.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt-1.10.1-1.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml-4.6.5-3.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe-1.1.1-12.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools-8.12.0-2.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack-1.0.3-2.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort-7.1.1-5.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-1:1.23.5-2.el9_7.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging-20.9-5.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply-3.11-14.el9.0.1.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend-3.1.0-2.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf-3.14.0-17.el9_7.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL-21.0.0-1.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-0.4.8-7.el9_7.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules-0.4.8-7.el9_7.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser-2.20-6.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing-2.4.7-9.el9.0.1.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-rados-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-rbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru-0.7-16.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-2.25.1-10.el9_6.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib-1.3.0-12.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-rgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes-2.5.1-5.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa-4.9-2.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml-1.16.0-1.el9.noarch 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy-1.9.3-2.el9.x86_64 2026-04-17T12:19:11.553 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora-5.0.0-2.el9.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml-0.10.2-6.el9.0.1.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions-4.15.0-1.el9.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3-1.26.5-6.el9_7.1.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client-1.2.3-2.el9.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec-1.3.13-1.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmltodict-0.12.0-15.el9.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile-2.0-10.el9.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-24.09.0-1.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service-24.09.0-1.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs-1.3.1-1.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: rbd-fuse-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: rbd-nbd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: re2-1:20211101-20.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: s3cmd-2.4.0-1.el9.noarch 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools-1:7.2-9.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: thrift-0.15.0-4.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-1.2.29-13.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl-1.2.29-13.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:11.554 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-04-17T12:19:11.665 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:11.665 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-04-17T12:19:11.665 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-04-17T12:19:11.665 INFO:teuthology.task.install:Skipping version verification because we have custom repos... 2026-04-17T12:19:11.665 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-04-17T12:19:11.665 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:11.665 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-04-17T12:19:11.698 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:11.699 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-04-17T12:19:11.739 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:11.740 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-04-17T12:19:11.771 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-04-17T12:19:11.771 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:11.771 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-04-17T12:19:11.806 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-04-17T12:19:11.877 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:11.878 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/daemon-helper 2026-04-17T12:19:11.907 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-04-17T12:19:11.976 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:11.976 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/daemon-helper 2026-04-17T12:19:12.001 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-04-17T12:19:12.070 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-04-17T12:19:12.071 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:12.071 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-04-17T12:19:12.101 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-04-17T12:19:12.171 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:12.171 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/adjust-ulimits 2026-04-17T12:19:12.199 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-04-17T12:19:12.264 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:12.265 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/adjust-ulimits 2026-04-17T12:19:12.295 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-04-17T12:19:12.362 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-04-17T12:19:12.363 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:12.363 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-04-17T12:19:12.390 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-04-17T12:19:12.458 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:12.459 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/usr/bin/stdin-killer 2026-04-17T12:19:12.487 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-04-17T12:19:12.557 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:12.557 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/usr/bin/stdin-killer 2026-04-17T12:19:12.593 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-04-17T12:19:12.667 INFO:teuthology.run_tasks:Running task ceph... 2026-04-17T12:19:12.719 INFO:tasks.ceph:Making ceph log dir writeable by non-root... 2026-04-17T12:19:12.719 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /var/log/ceph 2026-04-17T12:19:12.721 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /var/log/ceph 2026-04-17T12:19:12.723 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 777 /var/log/ceph 2026-04-17T12:19:12.759 INFO:tasks.ceph:Disabling ceph logrotate... 2026-04-17T12:19:12.759 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/logrotate.d/ceph 2026-04-17T12:19:12.795 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/logrotate.d/ceph 2026-04-17T12:19:12.802 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/logrotate.d/ceph 2026-04-17T12:19:12.834 INFO:tasks.ceph:Creating extra log directories... 2026-04-17T12:19:12.834 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-04-17T12:19:12.864 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-04-17T12:19:12.877 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/log/ceph/valgrind /var/log/ceph/profiling-logger 2026-04-17T12:19:12.912 INFO:tasks.ceph:Creating ceph cluster ceph... 2026-04-17T12:19:12.912 INFO:tasks.ceph:config {'conf': {'client': {'debug rgw': 20, 'debug rgw dedup': 20, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'global': {'osd_max_pg_log_entries': 10, 'osd_min_pg_log_entries': 10}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'bdev async discard': True, 'bdev enable discard': True, 'bluestore allocator': 'bitmap', 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug ms': 1, 'debug osd': 20, 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd mclock iops capacity threshold hdd': 49000, 'osd objectstore': 'bluestore', 'osd shutdown pgref assert': True}}, 'fs': 'xfs', 'mkfs_options': None, 'mount_options': None, 'skip_mgr_daemons': False, 'log_ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'cpu_profile': set(), 'cluster': 'ceph', 'mon_bind_msgr2': True, 'mon_bind_addrvec': True} 2026-04-17T12:19:12.912 INFO:tasks.ceph:ctx.config {'archive_path': '/archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589', 'branch': 'wip-sse-s3-on-v20.2.0', 'description': 'rgw/dedup/{beast bluestore-bitmap fixed-3-rgw ignore-pg-availability overrides supported-distros/{rocky_latest} tasks/{0-install test_dedup}}', 'email': None, 'first_in_suite': False, 'flavor': 'default', 'job_id': '5589', 'last_in_suite': False, 'machine_type': 'vps', 'name': 'supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps', 'no_nested_subset': False, 'openstack': [{'volumes': {'count': 4, 'size': 10}}], 'os_type': 'rocky', 'os_version': '9.7', 'overrides': {'admin_socket': {'branch': 'wip-sse-s3-on-v20.2.0'}, 'ansible.cephlab': {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}}, 'ceph': {'conf': {'client': {'debug rgw': 20, 'debug rgw dedup': 20, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'global': {'osd_max_pg_log_entries': 10, 'osd_min_pg_log_entries': 10}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'bdev async discard': True, 'bdev enable discard': True, 'bluestore allocator': 'bitmap', 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug ms': 1, 'debug osd': 20, 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd mclock iops capacity threshold hdd': 49000, 'osd objectstore': 'bluestore', 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'fs': 'xfs', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', '\\(PG_AVAILABILITY\\)', '\\(PG_DEGRADED\\)', '\\(POOL_APP_NOT_ENABLED\\)', 'not have an application enabled'], 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8'}, 'ceph-deploy': {'bluestore': True, 'conf': {'client': {'log file': '/var/log/ceph/ceph-$name.$pid.log'}, 'mon': {}, 'osd': {'bdev async discard': True, 'bdev enable discard': True, 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd objectstore': 'bluestore'}}, 'fs': 'xfs'}, 'cephadm': {'cephadm_binary_url': 'https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm', 'containers': {'image': 'harbor.clyso.com/custom-ceph/ceph/ceph:sse-s3-kmip-preview-not-for-production-7'}}, 'install': {'ceph': {'flavor': 'default', 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'repos': [{'name': 'ceph-source', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/SRPMS'}, {'name': 'ceph-noarch', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/noarch'}, {'name': 'ceph', 'priority': 1, 'url': 'https://s3.clyso.com/ces-packages/components/ceph-debug/rpm-20.2.0-21-gc03ba9ecf58/el9.clyso/x86_64'}]}, 'rgw': {'frontend': 'beast', 'storage classes': {'FROZEN': None, 'LUKEWARM': None}}, 's3tests': {'sha1': 'e0c4ff71baef6d5126a0201df5fe54196d89b296'}, 'selinux': {'allowlist': ['scontext=system_u:system_r:getty_t:s0']}, 'thrashosds': {'bdev_inject_crash': 2, 'bdev_inject_crash_probability': 0.5}, 'workunit': {'branch': 'tt-wip-sse-s3-on-v20.2.0', 'sha1': '909b66e106532fd1f1a49171c3f2eb7a193a6d0b'}}, 'owner': 'supriti', 'priority': 1000, 'repo': 'https://github.com/ceph/ceph.git', 'roles': [['mon.a', 'mon.c', 'mgr.y', 'osd.0', 'osd.1', 'osd.2', 'osd.3', 'client.0'], ['mon.b', 'mgr.x', 'osd.4', 'osd.5', 'osd.6', 'osd.7', 'client.1'], ['client.2']], 'seed': 7685, 'sha1': 'c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8', 'sleep_before_teardown': 0, 'suite': 'rgw', 'suite_branch': 'tt-wip-sse-s3-on-v20.2.0', 'suite_path': '/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa', 'suite_relpath': 'qa', 'suite_repo': 'http://git.local/ceph.git', 'suite_sha1': '909b66e106532fd1f1a49171c3f2eb7a193a6d0b', 'targets': {'vm02.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBA3v6WWmK1n6Sfq3xIlK6cW2X2vPTSyNwHkSmjaFBSTp1tVknwZUk43AfMuvt2UVUlvhDs84IZdlA3iLH+lTTbw=', 'vm06.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDJh0VKrTxfmhFAF2A4azvEw3oyVDAGgxyD40VIJFd/pNOEXihGKEUMFa6EeCcdnot3hUQvEpW9Wu96wInC9a3E=', 'vm08.local': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPtGs0s6hOkJKgF3AgFL0vCGkBCbq1q7LCFPDOhRpiT9seQ30oUK47pB0q4yI8pgBHC3VWBkTQLRETYTQROZIp4='}, 'tasks': [{'internal.save_config': None}, {'internal.check_lock': None}, {'internal.add_remotes': None}, {'console_log': None}, {'internal.connect': None}, {'internal.push_inventory': None}, {'internal.serialize_remote_roles': None}, {'internal.check_conflict': None}, {'internal.check_ceph_data': None}, {'internal.vm_setup': None}, {'internal.base': None}, {'internal.archive_upload': None}, {'internal.archive': None}, {'internal.coredump': None}, {'internal.sudo': None}, {'internal.syslog': None}, {'internal.timer': None}, {'pcp': None}, {'selinux': None}, {'ansible.cephlab': None}, {'clock': None}, {'install': None}, {'ceph': None}, {'openssl_keys': None}, {'rgw': ['client.0', 'client.1', 'client.2']}, {'tox': ['client.0']}, {'tox': ['client.0']}, {'dedup-tests': {'client.0': {'rgw_server': 'client.0'}}}], 'teuthology': {'fragments_dropped': [], 'meta': {}, 'postmerge': []}, 'teuthology_branch': 'clyso-debian-13', 'teuthology_repo': 'https://github.com/kshtsk/teuthology', 'teuthology_sha1': '1c580df7a9c7c2aadc272da296344fd99f27c444', 'timestamp': '2026-04-17_12:11:56', 'tube': 'vps', 'user': 'supriti', 'verbose': False, 'worker_log': '/home/teuthos/.teuthology/dispatcher/dispatcher.vps.3072398'} 2026-04-17T12:19:12.912 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/ceph.data 2026-04-17T12:19:12.937 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/ceph.data 2026-04-17T12:19:12.946 DEBUG:teuthology.orchestra.run.vm08:> install -d -m0755 -- /home/ubuntu/cephtest/ceph.data 2026-04-17T12:19:12.973 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m0777 -- /var/run/ceph 2026-04-17T12:19:12.998 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m0777 -- /var/run/ceph 2026-04-17T12:19:13.007 DEBUG:teuthology.orchestra.run.vm08:> sudo install -d -m0777 -- /var/run/ceph 2026-04-17T12:19:13.046 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:13.046 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-04-17T12:19:13.089 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-04-17T12:19:13.089 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_1 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 1043 Links: 1 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-04-17 12:19:02.551978524 +0000 2026-04-17T12:19:13.150 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-04-17 12:18:38.988956021 +0000 2026-04-17T12:19:13.151 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-04-17 12:18:38.988956021 +0000 2026-04-17T12:19:13.151 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-04-17 12:18:38.988956021 +0000 2026-04-17T12:19:13.151 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-04-17T12:19:13.223 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-04-17T12:19:13.223 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-04-17T12:19:13.223 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000365945 s, 1.4 MB/s 2026-04-17T12:19:13.225 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-04-17T12:19:13.288 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_2 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 1047 Links: 1 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-04-17 12:19:02.551978524 +0000 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-04-17 12:18:38.992956025 +0000 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-04-17 12:18:38.992956025 +0000 2026-04-17T12:19:13.350 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-04-17 12:18:38.992956025 +0000 2026-04-17T12:19:13.350 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-04-17T12:19:13.415 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-04-17T12:19:13.415 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-04-17T12:19:13.415 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000233527 s, 2.2 MB/s 2026-04-17T12:19:13.416 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-04-17T12:19:13.477 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_3 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 1059 Links: 1 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-04-17 12:19:02.552978525 +0000 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-04-17 12:18:38.993956026 +0000 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-04-17 12:18:38.993956026 +0000 2026-04-17T12:19:13.541 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-04-17 12:18:38.993956026 +0000 2026-04-17T12:19:13.541 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-04-17T12:19:13.611 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-04-17T12:19:13.611 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-04-17T12:19:13.611 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.00020302 s, 2.5 MB/s 2026-04-17T12:19:13.612 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-04-17T12:19:13.679 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_4 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 1061 Links: 1 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-04-17 12:19:02.552978525 +0000 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-04-17 12:18:38.997956030 +0000 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-04-17 12:18:38.997956030 +0000 2026-04-17T12:19:13.740 INFO:teuthology.orchestra.run.vm02.stdout: Birth: 2026-04-17 12:18:38.997956030 +0000 2026-04-17T12:19:13.740 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-04-17T12:19:13.812 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-04-17T12:19:13.812 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-04-17T12:19:13.812 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000279783 s, 1.8 MB/s 2026-04-17T12:19:13.813 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-04-17T12:19:13.875 INFO:tasks.ceph:osd dev map: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:13.875 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:13.875 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-04-17T12:19:13.898 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-04-17T12:19:13.898 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_1 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1059 Links: 1 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-17 12:19:07.833451064 +0000 2026-04-17T12:19:13.956 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-17 12:18:44.175423623 +0000 2026-04-17T12:19:13.957 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-17 12:18:44.175423623 +0000 2026-04-17T12:19:13.957 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-17 12:18:44.175423623 +0000 2026-04-17T12:19:13.957 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-04-17T12:19:14.027 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-17T12:19:14.027 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-17T12:19:14.027 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000132448 s, 3.9 MB/s 2026-04-17T12:19:14.028 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-04-17T12:19:14.092 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_2 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1049 Links: 1 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-17 12:19:07.833451064 +0000 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-17 12:18:44.170423617 +0000 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-17 12:18:44.170423617 +0000 2026-04-17T12:19:14.155 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-17 12:18:44.170423617 +0000 2026-04-17T12:19:14.156 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-04-17T12:19:14.223 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-17T12:19:14.229 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-17T12:19:14.229 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000153017 s, 3.3 MB/s 2026-04-17T12:19:14.230 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-04-17T12:19:14.293 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_3 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1046 Links: 1 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-17 12:19:07.833451064 +0000 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-17 12:18:44.169423616 +0000 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-17 12:18:44.169423616 +0000 2026-04-17T12:19:14.354 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-17 12:18:44.169423616 +0000 2026-04-17T12:19:14.354 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-04-17T12:19:14.420 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-17T12:19:14.420 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-17T12:19:14.420 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000190958 s, 2.7 MB/s 2026-04-17T12:19:14.421 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-04-17T12:19:14.484 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vg_nvme/lv_4 2026-04-17T12:19:14.545 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout:Device: 5h/5d Inode: 1063 Links: 1 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:device_t:s0 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-04-17 12:19:07.834451065 +0000 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-04-17 12:18:44.180423629 +0000 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-04-17 12:18:44.180423629 +0000 2026-04-17T12:19:14.546 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-04-17 12:18:44.180423629 +0000 2026-04-17T12:19:14.546 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-04-17T12:19:14.616 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-04-17T12:19:14.616 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-04-17T12:19:14.616 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000187813 s, 2.7 MB/s 2026-04-17T12:19:14.617 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-04-17T12:19:14.680 INFO:tasks.ceph:osd dev map: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:14.680 INFO:tasks.ceph:remote_to_roles_to_devs: {Remote(name='ubuntu@vm02.local'): {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'}, Remote(name='ubuntu@vm06.local'): {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'}} 2026-04-17T12:19:14.680 INFO:tasks.ceph:Generating config... 2026-04-17T12:19:14.681 INFO:tasks.ceph:[client] debug rgw = 20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[client] debug rgw dedup = 20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[client] setgroup = ceph 2026-04-17T12:19:14.681 INFO:tasks.ceph:[client] setuser = ceph 2026-04-17T12:19:14.681 INFO:tasks.ceph:[global] osd_max_pg_log_entries = 10 2026-04-17T12:19:14.681 INFO:tasks.ceph:[global] osd_min_pg_log_entries = 10 2026-04-17T12:19:14.681 INFO:tasks.ceph:[mgr] debug mgr = 20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[mgr] debug ms = 1 2026-04-17T12:19:14.681 INFO:tasks.ceph:[mon] debug mon = 20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[mon] debug ms = 1 2026-04-17T12:19:14.681 INFO:tasks.ceph:[mon] debug paxos = 20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] bdev async discard = True 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] bdev enable discard = True 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] bluestore allocator = bitmap 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] bluestore block size = 96636764160 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] bluestore fsck on mount = True 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] debug bluefs = 1/20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] debug bluestore = 1/20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] debug ms = 1 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] debug osd = 20 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] debug rocksdb = 4/10 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] mon osd backfillfull_ratio = 0.85 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] mon osd full ratio = 0.9 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] mon osd nearfull ratio = 0.8 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] osd failsafe full ratio = 0.95 2026-04-17T12:19:14.681 INFO:tasks.ceph:[osd] osd mclock iops capacity threshold hdd = 49000 2026-04-17T12:19:14.682 INFO:tasks.ceph:[osd] osd objectstore = bluestore 2026-04-17T12:19:14.682 INFO:tasks.ceph:[osd] osd shutdown pgref assert = True 2026-04-17T12:19:14.682 INFO:tasks.ceph:Setting up mon.a... 2026-04-17T12:19:14.682 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring /etc/ceph/ceph.keyring 2026-04-17T12:19:14.730 INFO:teuthology.orchestra.run.vm02.stdout:creating /etc/ceph/ceph.keyring 2026-04-17T12:19:14.734 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=mon. /etc/ceph/ceph.keyring 2026-04-17T12:19:14.782 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-17T12:19:14.809 DEBUG:tasks.ceph:Ceph mon addresses: [('mon.a', '192.168.123.102'), ('mon.c', '[v2:192.168.123.102:3301,v1:192.168.123.102:6790]'), ('mon.b', '192.168.123.106')] 2026-04-17T12:19:14.809 DEBUG:tasks.ceph:writing out conf {'global': {'chdir': '', 'pid file': '/var/run/ceph/$cluster-$name.pid', 'auth supported': 'cephx', 'filestore xattr use omap': 'true', 'mon clock drift allowed': '1.000', 'osd crush chooseleaf type': '0', 'auth debug': 'true', 'ms die on old message': 'true', 'ms die on bug': 'true', 'mon max pg per osd': '10000', 'mon pg warn max object skew': '0', 'osd_pool_default_pg_autoscale_mode': 'off', 'osd pool default size': '2', 'mon osd allow primary affinity': 'true', 'mon osd allow pg remap': 'true', 'mon warn on legacy crush tunables': 'false', 'mon warn on crush straw calc version zero': 'false', 'mon warn on no sortbitwise': 'false', 'mon warn on osd down out interval zero': 'false', 'mon warn on too few osds': 'false', 'mon_warn_on_pool_pg_num_not_power_of_two': 'false', 'mon_warn_on_pool_no_redundancy': 'false', 'mon_allow_pool_size_one': 'true', 'osd pool default erasure code profile': 'plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd', 'osd default data pool replay window': '5', 'mon allow pool delete': 'true', 'mon cluster log file level': 'debug', 'debug asserts on shutdown': 'true', 'mon health detail to clog': 'false', 'mon host': '192.168.123.102,[v2:192.168.123.102:3301,v1:192.168.123.102:6790],192.168.123.106', 'osd_max_pg_log_entries': 10, 'osd_min_pg_log_entries': 10}, 'osd': {'osd journal size': '100', 'osd scrub load threshold': '5.0', 'osd scrub max interval': '600', 'osd mclock profile': 'high_recovery_ops', 'osd recover clone overlap': 'true', 'osd recovery max chunk': '1048576', 'osd debug shutdown': 'true', 'osd debug op order': 'true', 'osd debug verify stray on activate': 'true', 'osd debug trim objects': 'true', 'osd open classes on start': 'true', 'osd debug pg log writeout': 'true', 'osd deep scrub update digest min age': '30', 'osd map max advance': '10', 'journal zero on create': 'true', 'filestore ondisk finisher threads': '3', 'filestore apply finisher threads': '3', 'bdev debug aio': 'true', 'osd debug misdirected ops': 'true', 'bdev async discard': True, 'bdev enable discard': True, 'bluestore allocator': 'bitmap', 'bluestore block size': 96636764160, 'bluestore fsck on mount': True, 'debug bluefs': '1/20', 'debug bluestore': '1/20', 'debug ms': 1, 'debug osd': 20, 'debug rocksdb': '4/10', 'mon osd backfillfull_ratio': 0.85, 'mon osd full ratio': 0.9, 'mon osd nearfull ratio': 0.8, 'osd failsafe full ratio': 0.95, 'osd mclock iops capacity threshold hdd': 49000, 'osd objectstore': 'bluestore', 'osd shutdown pgref assert': True}, 'mgr': {'debug ms': 1, 'debug mgr': 20, 'debug mon': '20', 'debug auth': '20', 'mon reweight min pgs per osd': '4', 'mon reweight min bytes per osd': '10', 'mgr/telemetry/nag': 'false'}, 'mon': {'debug ms': 1, 'debug mon': 20, 'debug paxos': 20, 'debug auth': '20', 'mon data avail warn': '5', 'mon mgr mkfs grace': '240', 'mon reweight min pgs per osd': '4', 'mon osd reporter subtree level': 'osd', 'mon osd prime pg temp': 'true', 'mon reweight min bytes per osd': '10', 'auth mon ticket ttl': '660', 'auth service ticket ttl': '240', 'mon_warn_on_insecure_global_id_reclaim': 'false', 'mon_warn_on_insecure_global_id_reclaim_allowed': 'false', 'mon_down_mkfs_grace': '2m', 'mon_warn_on_filestore_osds': 'false'}, 'client': {'rgw cache enabled': 'true', 'rgw enable ops log': 'true', 'rgw enable usage log': 'true', 'log file': '/var/log/ceph/$cluster-$name.$pid.log', 'admin socket': '/var/run/ceph/$cluster-$name.$pid.asok', 'debug rgw': 20, 'debug rgw dedup': 20, 'setgroup': 'ceph', 'setuser': 'ceph'}, 'mon.a': {}, 'mon.c': {}, 'mon.b': {}} 2026-04-17T12:19:14.809 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:14.809 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/ceph.tmp.conf 2026-04-17T12:19:14.868 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage monmaptool -c /home/ubuntu/cephtest/ceph.tmp.conf --create --clobber --enable-all-features --add a 192.168.123.102 --addv c '[v2:192.168.123.102:3301,v1:192.168.123.102:6790]' --add b 192.168.123.106 --print /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:14.948 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:19:14.948 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool: monmap file /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool: generated fsid 50e117d9-5cfb-424b-9f73-1b1dada394f8 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = tentacle 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:epoch 0 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:fsid 50e117d9-5cfb-424b-9f73-1b1dada394f8 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:last_changed 2026-04-17T12:19:14.948804+0000 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:created 2026-04-17T12:19:14.948804+0000 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:min_mon_release 20 (tentacle) 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:election_strategy: 1 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:1: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.b 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:2: [v2:192.168.123.102:3301/0,v1:192.168.123.102:6790/0] mon.c 2026-04-17T12:19:14.949 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool: writing epoch 0 to /home/ubuntu/cephtest/ceph.monmap (3 monitors) 2026-04-17T12:19:14.950 DEBUG:teuthology.orchestra.run.vm02:> rm -- /home/ubuntu/cephtest/ceph.tmp.conf 2026-04-17T12:19:15.009 INFO:tasks.ceph:Writing /etc/ceph/ceph.conf for FSID 50e117d9-5cfb-424b-9f73-1b1dada394f8... 2026-04-17T12:19:15.010 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/ceph.conf && sudo chmod 0644 /etc/ceph/ceph.conf > /dev/null 2026-04-17T12:19:15.052 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/ceph.conf && sudo chmod 0644 /etc/ceph/ceph.conf > /dev/null 2026-04-17T12:19:15.055 DEBUG:teuthology.orchestra.run.vm08:> sudo mkdir -p /etc/ceph && sudo chmod 0755 /etc/ceph && sudo tee /etc/ceph/ceph.conf && sudo chmod 0644 /etc/ceph/ceph.conf > /dev/null 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: chdir = "" 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: auth supported = cephx 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: filestore xattr use omap = true 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: mon clock drift allowed = 1.000 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: osd crush chooseleaf type = 0 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: auth debug = true 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: ms die on old message = true 2026-04-17T12:19:15.102 INFO:teuthology.orchestra.run.vm06.stdout: ms die on bug = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon max pg per osd = 10000 # >= luminous 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon pg warn max object skew = 0 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: # disable pg_autoscaler by default for new pools 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd_pool_default_pg_autoscale_mode = off 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default size = 2 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow primary affinity = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon osd allow pg remap = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on legacy crush tunables = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on crush straw calc version zero = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on no sortbitwise = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on osd down out interval zero = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon warn on too few osds = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_pool_no_redundancy = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon_allow_pool_size_one = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd default data pool replay window = 5 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon allow pool delete = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon cluster log file level = debug 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: debug asserts on shutdown = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon health detail to clog = false 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: mon host = "192.168.123.102,[v2:192.168.123.102:3301,v1:192.168.123.102:6790],192.168.123.106" 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd_max_pg_log_entries = 10 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd_min_pg_log_entries = 10 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: fsid = 50e117d9-5cfb-424b-9f73-1b1dada394f8 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout:[osd] 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd journal size = 100 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub load threshold = 5.0 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd scrub max interval = 600 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock profile = high_recovery_ops 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd recover clone overlap = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd recovery max chunk = 1048576 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd debug shutdown = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd debug op order = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd debug verify stray on activate = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: osd debug trim objects = true 2026-04-17T12:19:15.103 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd open classes on start = true 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd debug pg log writeout = true 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd deep scrub update digest min age = 30 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd map max advance = 10 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: journal zero on create = true 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: filestore ondisk finisher threads = 3 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: filestore apply finisher threads = 3 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: bdev debug aio = true 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd debug misdirected ops = true 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: bdev async discard = True 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: bdev enable discard = True 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: bluestore allocator = bitmap 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: bluestore block size = 96636764160 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: bluestore fsck on mount = True 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug bluefs = 1/20 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug bluestore = 1/20 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug osd = 20 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug rocksdb = 4/10 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: mon osd backfillfull_ratio = 0.85 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: mon osd full ratio = 0.9 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: mon osd nearfull ratio = 0.8 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd failsafe full ratio = 0.95 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd mclock iops capacity threshold hdd = 49000 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd objectstore = bluestore 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: osd shutdown pgref assert = True 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout:[mgr] 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug mgr = 20 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: mgr/telemetry/nag = false 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout:[mon] 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug ms = 1 2026-04-17T12:19:15.104 INFO:teuthology.orchestra.run.vm06.stdout: debug mon = 20 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: debug paxos = 20 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: debug auth = 20 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon data avail warn = 5 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon mgr mkfs grace = 240 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min pgs per osd = 4 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon osd reporter subtree level = osd 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon osd prime pg temp = true 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon reweight min bytes per osd = 10 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: auth mon ticket ttl = 660 # 11m 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: auth service ticket ttl = 240 # 4m 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: # don't complain about insecure global_id in the test suite 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: # 1m isn't quite enough 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon_down_mkfs_grace = 2m 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: mon_warn_on_filestore_osds = false 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout:[client] 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: rgw cache enabled = true 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable ops log = true 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: rgw enable usage log = true 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw = 20 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: debug rgw dedup = 20 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: setgroup = ceph 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout: setuser = ceph 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout:[mon.a] 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout:[mon.c] 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm06.stdout:[mon.b] 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm02.stdout: chdir = "" 2026-04-17T12:19:15.105 INFO:teuthology.orchestra.run.vm02.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: auth supported = cephx 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: filestore xattr use omap = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon clock drift allowed = 1.000 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: osd crush chooseleaf type = 0 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: auth debug = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: ms die on old message = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: ms die on bug = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon max pg per osd = 10000 # >= luminous 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon pg warn max object skew = 0 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: # disable pg_autoscaler by default for new pools 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: osd_pool_default_pg_autoscale_mode = off 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: osd pool default size = 2 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon osd allow primary affinity = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon osd allow pg remap = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on legacy crush tunables = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on crush straw calc version zero = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on no sortbitwise = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on osd down out interval zero = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon warn on too few osds = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_pool_no_redundancy = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon_allow_pool_size_one = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: osd default data pool replay window = 5 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon allow pool delete = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon cluster log file level = debug 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: debug asserts on shutdown = true 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon health detail to clog = false 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: mon host = "192.168.123.102,[v2:192.168.123.102:3301,v1:192.168.123.102:6790],192.168.123.106" 2026-04-17T12:19:15.106 INFO:teuthology.orchestra.run.vm02.stdout: osd_max_pg_log_entries = 10 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd_min_pg_log_entries = 10 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: fsid = 50e117d9-5cfb-424b-9f73-1b1dada394f8 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout:[osd] 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd journal size = 100 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd scrub load threshold = 5.0 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd scrub max interval = 600 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock profile = high_recovery_ops 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd recover clone overlap = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd recovery max chunk = 1048576 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd debug shutdown = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd debug op order = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd debug verify stray on activate = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd debug trim objects = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd open classes on start = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd debug pg log writeout = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd deep scrub update digest min age = 30 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd map max advance = 10 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: journal zero on create = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: filestore ondisk finisher threads = 3 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: filestore apply finisher threads = 3 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: bdev debug aio = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd debug misdirected ops = true 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: bdev async discard = True 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: bdev enable discard = True 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: bluestore allocator = bitmap 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: bluestore block size = 96636764160 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: bluestore fsck on mount = True 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: debug bluefs = 1/20 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: debug bluestore = 1/20 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: debug osd = 20 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: debug rocksdb = 4/10 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: mon osd backfillfull_ratio = 0.85 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: mon osd full ratio = 0.9 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: mon osd nearfull ratio = 0.8 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd failsafe full ratio = 0.95 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd mclock iops capacity threshold hdd = 49000 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd objectstore = bluestore 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: osd shutdown pgref assert = True 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout:[mgr] 2026-04-17T12:19:15.107 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug mgr = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug mon = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug auth = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min pgs per osd = 4 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min bytes per osd = 10 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mgr/telemetry/nag = false 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout:[mon] 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug ms = 1 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug mon = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug paxos = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug auth = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon data avail warn = 5 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon mgr mkfs grace = 240 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min pgs per osd = 4 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon osd reporter subtree level = osd 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon osd prime pg temp = true 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon reweight min bytes per osd = 10 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: auth mon ticket ttl = 660 # 11m 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: auth service ticket ttl = 240 # 4m 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: # don't complain about insecure global_id in the test suite 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: # 1m isn't quite enough 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon_down_mkfs_grace = 2m 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: mon_warn_on_filestore_osds = false 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout:[client] 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: rgw cache enabled = true 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: rgw enable ops log = true 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: rgw enable usage log = true 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: debug rgw dedup = 20 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: setgroup = ceph 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout: setuser = ceph 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout:[mon.a] 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout:[mon.c] 2026-04-17T12:19:15.108 INFO:teuthology.orchestra.run.vm02.stdout:[mon.b] 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout:[global] 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: chdir = "" 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: pid file = /var/run/ceph/$cluster-$name.pid 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: auth supported = cephx 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: filestore xattr use omap = true 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: mon clock drift allowed = 1.000 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: osd crush chooseleaf type = 0 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: auth debug = true 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: ms die on old message = true 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: ms die on bug = true 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: mon max pg per osd = 10000 # >= luminous 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: mon pg warn max object skew = 0 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.110 INFO:teuthology.orchestra.run.vm08.stdout: # disable pg_autoscaler by default for new pools 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd_pool_default_pg_autoscale_mode = off 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default size = 2 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow primary affinity = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon osd allow pg remap = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on legacy crush tunables = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on crush straw calc version zero = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on no sortbitwise = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on osd down out interval zero = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon warn on too few osds = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_pg_num_not_power_of_two = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_pool_no_redundancy = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon_allow_pool_size_one = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd default data pool replay window = 5 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon allow pool delete = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon cluster log file level = debug 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: debug asserts on shutdown = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon health detail to clog = false 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: mon host = "192.168.123.102,[v2:192.168.123.102:3301,v1:192.168.123.102:6790],192.168.123.106" 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd_max_pg_log_entries = 10 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd_min_pg_log_entries = 10 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: fsid = 50e117d9-5cfb-424b-9f73-1b1dada394f8 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout:[osd] 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd journal size = 100 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub load threshold = 5.0 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd scrub max interval = 600 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock profile = high_recovery_ops 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd recover clone overlap = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd recovery max chunk = 1048576 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd debug shutdown = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd debug op order = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd debug verify stray on activate = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd debug trim objects = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd open classes on start = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd debug pg log writeout = true 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd deep scrub update digest min age = 30 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.111 INFO:teuthology.orchestra.run.vm08.stdout: osd map max advance = 10 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: journal zero on create = true 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: filestore ondisk finisher threads = 3 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: filestore apply finisher threads = 3 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: bdev debug aio = true 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: osd debug misdirected ops = true 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: bdev async discard = True 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: bdev enable discard = True 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: bluestore allocator = bitmap 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: bluestore block size = 96636764160 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: bluestore fsck on mount = True 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug bluefs = 1/20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug bluestore = 1/20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug osd = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug rocksdb = 4/10 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon osd backfillfull_ratio = 0.85 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon osd full ratio = 0.9 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon osd nearfull ratio = 0.8 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: osd failsafe full ratio = 0.95 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: osd mclock iops capacity threshold hdd = 49000 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: osd objectstore = bluestore 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: osd shutdown pgref assert = True 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout:[mgr] 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug mgr = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mgr/telemetry/nag = false 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout:[mon] 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug ms = 1 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug mon = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug paxos = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: debug auth = 20 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon data avail warn = 5 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon mgr mkfs grace = 240 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min pgs per osd = 4 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon osd reporter subtree level = osd 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon osd prime pg temp = true 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: mon reweight min bytes per osd = 10 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: # rotate auth tickets quickly to exercise renewal paths 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: auth mon ticket ttl = 660 # 11m 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: auth service ticket ttl = 240 # 4m 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.112 INFO:teuthology.orchestra.run.vm08.stdout: # don't complain about insecure global_id in the test suite 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim = false 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_insecure_global_id_reclaim_allowed = false 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: # 1m isn't quite enough 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: mon_down_mkfs_grace = 2m 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: mon_warn_on_filestore_osds = false 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout:[client] 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: rgw cache enabled = true 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable ops log = true 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: rgw enable usage log = true 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: log file = /var/log/ceph/$cluster-$name.$pid.log 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: admin socket = /var/run/ceph/$cluster-$name.$pid.asok 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw = 20 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: debug rgw dedup = 20 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: setgroup = ceph 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout: setuser = ceph 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout:[mon.a] 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout:[mon.c] 2026-04-17T12:19:15.113 INFO:teuthology.orchestra.run.vm08.stdout:[mon.b] 2026-04-17T12:19:15.123 INFO:tasks.ceph:Creating admin key on mon.a... 2026-04-17T12:19:15.123 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --gen-key --name=client.admin --cap mon 'allow *' --cap osd 'allow *' --cap mds 'allow *' --cap mgr 'allow *' /etc/ceph/ceph.keyring 2026-04-17T12:19:15.178 INFO:tasks.ceph:Copying monmap to all nodes... 2026-04-17T12:19:15.178 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:15.178 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.keyring of=/dev/stdout 2026-04-17T12:19:15.198 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:15.198 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/ceph.monmap of=/dev/stdout 2026-04-17T12:19:15.256 INFO:tasks.ceph:Sending monmap to node ubuntu@vm02.local 2026-04-17T12:19:15.256 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:15.256 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.keyring 2026-04-17T12:19:15.256 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-17T12:19:15.339 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:15.339 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:15.401 INFO:tasks.ceph:Sending monmap to node ubuntu@vm06.local 2026-04-17T12:19:15.401 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:15.401 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.keyring 2026-04-17T12:19:15.401 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-17T12:19:15.445 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:15.446 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:15.505 INFO:tasks.ceph:Sending monmap to node ubuntu@vm08.local 2026-04-17T12:19:15.505 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:15.505 DEBUG:teuthology.orchestra.run.vm08:> sudo dd of=/etc/ceph/ceph.keyring 2026-04-17T12:19:15.505 DEBUG:teuthology.orchestra.run.vm08:> sudo chmod 0644 /etc/ceph/ceph.keyring 2026-04-17T12:19:15.547 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:15.547 DEBUG:teuthology.orchestra.run.vm08:> dd of=/home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:15.608 INFO:tasks.ceph:Setting up mon nodes... 2026-04-17T12:19:15.608 INFO:tasks.ceph:Setting up mgr nodes... 2026-04-17T12:19:15.608 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/mgr/ceph-y && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=mgr.y /var/lib/ceph/mgr/ceph-y/keyring 2026-04-17T12:19:15.664 INFO:teuthology.orchestra.run.vm02.stdout:creating /var/lib/ceph/mgr/ceph-y/keyring 2026-04-17T12:19:15.667 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/mgr/ceph-x && sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=mgr.x /var/lib/ceph/mgr/ceph-x/keyring 2026-04-17T12:19:15.730 INFO:teuthology.orchestra.run.vm06.stdout:creating /var/lib/ceph/mgr/ceph-x/keyring 2026-04-17T12:19:15.734 INFO:tasks.ceph:Setting up mds nodes... 2026-04-17T12:19:15.734 INFO:tasks.ceph_client:Setting up client nodes... 2026-04-17T12:19:15.734 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.0 /etc/ceph/ceph.client.0.keyring && sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-04-17T12:19:15.776 INFO:teuthology.orchestra.run.vm02.stdout:creating /etc/ceph/ceph.client.0.keyring 2026-04-17T12:19:15.791 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.1 /etc/ceph/ceph.client.1.keyring && sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-04-17T12:19:15.835 INFO:teuthology.orchestra.run.vm06.stdout:creating /etc/ceph/ceph.client.1.keyring 2026-04-17T12:19:15.848 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool --create-keyring --gen-key --name=client.2 /etc/ceph/ceph.client.2.keyring && sudo chmod 0644 /etc/ceph/ceph.client.2.keyring 2026-04-17T12:19:15.898 INFO:teuthology.orchestra.run.vm08.stdout:creating /etc/ceph/ceph.client.2.keyring 2026-04-17T12:19:15.910 INFO:tasks.ceph:Running mkfs on osd nodes... 2026-04-17T12:19:15.910 INFO:tasks.ceph:ctx.disk_config.remote_to_roles_to_dev: {Remote(name='ubuntu@vm02.local'): {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'}, Remote(name='ubuntu@vm06.local'): {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'}} 2026-04-17T12:19:15.910 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/ceph-0 2026-04-17T12:19:15.944 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:15.944 INFO:tasks.ceph:role: osd.0 2026-04-17T12:19:15.944 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm02.local 2026-04-17T12:19:15.944 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:16.013 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:16.014 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:16.014 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:16.020 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-04-17T12:19:16.031 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm02.local -o noatime 2026-04-17T12:19:16.031 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/ceph-0 2026-04-17T12:19:16.106 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-0 2026-04-17T12:19:16.176 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/ceph-1 2026-04-17T12:19:16.242 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:16.242 INFO:tasks.ceph:role: osd.1 2026-04-17T12:19:16.242 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm02.local 2026-04-17T12:19:16.242 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:16.308 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:16.313 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-04-17T12:19:16.314 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm02.local -o noatime 2026-04-17T12:19:16.314 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/ceph-1 2026-04-17T12:19:16.390 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-1 2026-04-17T12:19:16.464 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/ceph-2 2026-04-17T12:19:16.538 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:16.538 INFO:tasks.ceph:role: osd.2 2026-04-17T12:19:16.538 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm02.local 2026-04-17T12:19:16.538 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:16.606 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:16.611 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-04-17T12:19:16.613 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm02.local -o noatime 2026-04-17T12:19:16.613 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/ceph-2 2026-04-17T12:19:16.690 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-2 2026-04-17T12:19:16.760 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/osd/ceph-3 2026-04-17T12:19:16.828 INFO:tasks.ceph:roles_to_devs: {'osd.0': '/dev/vg_nvme/lv_1', 'osd.1': '/dev/vg_nvme/lv_2', 'osd.2': '/dev/vg_nvme/lv_3', 'osd.3': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:16.828 INFO:tasks.ceph:role: osd.3 2026-04-17T12:19:16.828 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_4 on ubuntu@vm02.local 2026-04-17T12:19:16.828 DEBUG:teuthology.orchestra.run.vm02:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_4 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout:meta-data=/dev/vg_nvme/lv_4 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:16.896 INFO:teuthology.orchestra.run.vm02.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:16.901 INFO:teuthology.orchestra.run.vm02.stdout:Discarding blocks...Done. 2026-04-17T12:19:16.904 INFO:tasks.ceph:mount /dev/vg_nvme/lv_4 on ubuntu@vm02.local -o noatime 2026-04-17T12:19:16.904 DEBUG:teuthology.orchestra.run.vm02:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_4 /var/lib/ceph/osd/ceph-3 2026-04-17T12:19:16.982 DEBUG:teuthology.orchestra.run.vm02:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-3 2026-04-17T12:19:17.062 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 0 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:17.158 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:17.156+0000 7fd043bc7900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-0/keyring: can't open /var/lib/ceph/osd/ceph-0/keyring: (2) No such file or directory 2026-04-17T12:19:17.158 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:17.156+0000 7fd043bc7900 -1 created new key in keyring /var/lib/ceph/osd/ceph-0/keyring 2026-04-17T12:19:17.158 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:17.156+0000 7fd043bc7900 -1 bdev(0x55662b0e3800 /var/lib/ceph/osd/ceph-0/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:17.158 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:17.156+0000 7fd043bc7900 -1 bluestore(/var/lib/ceph/osd/ceph-0) _read_fsid unparsable uuid 2026-04-17T12:19:17.884 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 2026-04-17T12:19:17.917 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 1 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:18.010 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.007+0000 7fab95ddf900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-1/keyring: can't open /var/lib/ceph/osd/ceph-1/keyring: (2) No such file or directory 2026-04-17T12:19:18.010 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.007+0000 7fab95ddf900 -1 created new key in keyring /var/lib/ceph/osd/ceph-1/keyring 2026-04-17T12:19:18.010 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.007+0000 7fab95ddf900 -1 bdev(0x55d81a801800 /var/lib/ceph/osd/ceph-1/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:18.010 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.008+0000 7fab95ddf900 -1 bluestore(/var/lib/ceph/osd/ceph-1) _read_fsid unparsable uuid 2026-04-17T12:19:18.781 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 2026-04-17T12:19:18.810 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 2 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:18.895 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.893+0000 7f916105a900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-2/keyring: can't open /var/lib/ceph/osd/ceph-2/keyring: (2) No such file or directory 2026-04-17T12:19:18.895 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.893+0000 7f916105a900 -1 created new key in keyring /var/lib/ceph/osd/ceph-2/keyring 2026-04-17T12:19:18.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.893+0000 7f916105a900 -1 bdev(0x55be6f05f800 /var/lib/ceph/osd/ceph-2/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:18.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:18.894+0000 7f916105a900 -1 bluestore(/var/lib/ceph/osd/ceph-2) _read_fsid unparsable uuid 2026-04-17T12:19:19.622 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-2 2026-04-17T12:19:19.656 DEBUG:teuthology.orchestra.run.vm02:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 3 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:19.740 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:19.738+0000 7f3e685db900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-3/keyring: can't open /var/lib/ceph/osd/ceph-3/keyring: (2) No such file or directory 2026-04-17T12:19:19.740 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:19.738+0000 7f3e685db900 -1 created new key in keyring /var/lib/ceph/osd/ceph-3/keyring 2026-04-17T12:19:19.740 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:19.738+0000 7f3e685db900 -1 bdev(0x563bba68b800 /var/lib/ceph/osd/ceph-3/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:19.740 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:19.738+0000 7f3e685db900 -1 bluestore(/var/lib/ceph/osd/ceph-3) _read_fsid unparsable uuid 2026-04-17T12:19:20.504 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-3 2026-04-17T12:19:20.571 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-4 2026-04-17T12:19:20.599 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:20.599 INFO:tasks.ceph:role: osd.4 2026-04-17T12:19:20.599 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_1 on ubuntu@vm06.local 2026-04-17T12:19:20.599 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_1 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_1 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:20.671 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:20.675 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-17T12:19:20.679 INFO:tasks.ceph:mount /dev/vg_nvme/lv_1 on ubuntu@vm06.local -o noatime 2026-04-17T12:19:20.679 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_1 /var/lib/ceph/osd/ceph-4 2026-04-17T12:19:20.760 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-4 2026-04-17T12:19:20.833 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-5 2026-04-17T12:19:20.901 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:20.902 INFO:tasks.ceph:role: osd.5 2026-04-17T12:19:20.902 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_2 on ubuntu@vm06.local 2026-04-17T12:19:20.902 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_2 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_2 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:20.968 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:20.973 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-17T12:19:20.975 INFO:tasks.ceph:mount /dev/vg_nvme/lv_2 on ubuntu@vm06.local -o noatime 2026-04-17T12:19:20.975 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_2 /var/lib/ceph/osd/ceph-5 2026-04-17T12:19:21.045 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-5 2026-04-17T12:19:21.119 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-6 2026-04-17T12:19:21.188 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:21.188 INFO:tasks.ceph:role: osd.6 2026-04-17T12:19:21.188 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_3 on ubuntu@vm06.local 2026-04-17T12:19:21.188 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_3 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_3 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:21.263 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:21.269 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-17T12:19:21.270 INFO:tasks.ceph:mount /dev/vg_nvme/lv_3 on ubuntu@vm06.local -o noatime 2026-04-17T12:19:21.270 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_3 /var/lib/ceph/osd/ceph-6 2026-04-17T12:19:21.344 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-6 2026-04-17T12:19:21.413 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/osd/ceph-7 2026-04-17T12:19:21.482 INFO:tasks.ceph:roles_to_devs: {'osd.4': '/dev/vg_nvme/lv_1', 'osd.5': '/dev/vg_nvme/lv_2', 'osd.6': '/dev/vg_nvme/lv_3', 'osd.7': '/dev/vg_nvme/lv_4'} 2026-04-17T12:19:21.482 INFO:tasks.ceph:role: osd.7 2026-04-17T12:19:21.482 INFO:tasks.ceph:['mkfs.xfs', '-f', '-i', 'size=2048'] on /dev/vg_nvme/lv_4 on ubuntu@vm06.local 2026-04-17T12:19:21.482 DEBUG:teuthology.orchestra.run.vm06:> yes | sudo mkfs.xfs -f -i size=2048 /dev/vg_nvme/lv_4 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout:meta-data=/dev/vg_nvme/lv_4 isize=2048 agcount=4, agsize=1310464 blks 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 attr=2, projid32bit=1 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout: = crc=1 finobt=1, sparse=1, rmapbt=0 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout: = reflink=1 bigtime=1 inobtcount=1 nrext64=0 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout:data = bsize=4096 blocks=5241856, imaxpct=25 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout: = sunit=0 swidth=0 blks 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout:naming =version 2 bsize=4096 ascii-ci=0, ftype=1 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout:log =internal log bsize=4096 blocks=16384, version=2 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout: = sectsz=512 sunit=0 blks, lazy-count=1 2026-04-17T12:19:21.551 INFO:teuthology.orchestra.run.vm06.stdout:realtime =none extsz=4096 blocks=0, rtextents=0 2026-04-17T12:19:21.588 INFO:teuthology.orchestra.run.vm06.stdout:Discarding blocks...Done. 2026-04-17T12:19:21.589 INFO:tasks.ceph:mount /dev/vg_nvme/lv_4 on ubuntu@vm06.local -o noatime 2026-04-17T12:19:21.589 DEBUG:teuthology.orchestra.run.vm06:> sudo mount -t xfs -o noatime /dev/vg_nvme/lv_4 /var/lib/ceph/osd/ceph-7 2026-04-17T12:19:21.661 DEBUG:teuthology.orchestra.run.vm06:> sudo /sbin/restorecon /var/lib/ceph/osd/ceph-7 2026-04-17T12:19:21.731 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 4 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:21.821 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:21.820+0000 7f4be53be900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-4/keyring: can't open /var/lib/ceph/osd/ceph-4/keyring: (2) No such file or directory 2026-04-17T12:19:21.821 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:21.820+0000 7f4be53be900 -1 created new key in keyring /var/lib/ceph/osd/ceph-4/keyring 2026-04-17T12:19:21.821 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:21.820+0000 7f4be53be900 -1 bdev(0x5588c2503800 /var/lib/ceph/osd/ceph-4/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:21.822 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:21.821+0000 7f4be53be900 -1 bluestore(/var/lib/ceph/osd/ceph-4) _read_fsid unparsable uuid 2026-04-17T12:19:22.465 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-4 2026-04-17T12:19:22.491 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 5 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:22.576 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:22.575+0000 7fea105db900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-5/keyring: can't open /var/lib/ceph/osd/ceph-5/keyring: (2) No such file or directory 2026-04-17T12:19:22.576 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:22.575+0000 7fea105db900 -1 created new key in keyring /var/lib/ceph/osd/ceph-5/keyring 2026-04-17T12:19:22.576 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:22.576+0000 7fea105db900 -1 bdev(0x561f2ccf1800 /var/lib/ceph/osd/ceph-5/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:22.576 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:22.576+0000 7fea105db900 -1 bluestore(/var/lib/ceph/osd/ceph-5) _read_fsid unparsable uuid 2026-04-17T12:19:23.430 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 2026-04-17T12:19:23.456 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 6 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:23.538 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:23.537+0000 7f39062ae900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-6/keyring: can't open /var/lib/ceph/osd/ceph-6/keyring: (2) No such file or directory 2026-04-17T12:19:23.538 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:23.537+0000 7f39062ae900 -1 created new key in keyring /var/lib/ceph/osd/ceph-6/keyring 2026-04-17T12:19:23.538 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:23.537+0000 7f39062ae900 -1 bdev(0x563b94399800 /var/lib/ceph/osd/ceph-6/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:23.538 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:23.537+0000 7f39062ae900 -1 bluestore(/var/lib/ceph/osd/ceph-6) _read_fsid unparsable uuid 2026-04-17T12:19:24.185 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 2026-04-17T12:19:24.253 DEBUG:teuthology.orchestra.run.vm06:> sudo MALLOC_CHECK_=3 adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-osd --no-mon-config --cluster ceph --mkfs --mkkey -i 7 --monmap /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:24.336 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:24.335+0000 7ff2a1b1b900 -1 auth: error reading file: /var/lib/ceph/osd/ceph-7/keyring: can't open /var/lib/ceph/osd/ceph-7/keyring: (2) No such file or directory 2026-04-17T12:19:24.337 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:24.335+0000 7ff2a1b1b900 -1 created new key in keyring /var/lib/ceph/osd/ceph-7/keyring 2026-04-17T12:19:24.337 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:24.335+0000 7ff2a1b1b900 -1 bdev(0x557419a6f800 /var/lib/ceph/osd/ceph-7/block) open stat got: (1) Operation not permitted 2026-04-17T12:19:24.337 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:19:24.336+0000 7ff2a1b1b900 -1 bluestore(/var/lib/ceph/osd/ceph-7) _read_fsid unparsable uuid 2026-04-17T12:19:25.006 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 2026-04-17T12:19:25.071 INFO:tasks.ceph:Reading keys from all nodes... 2026-04-17T12:19:25.071 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:25.071 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/mgr/ceph-y/keyring of=/dev/stdout 2026-04-17T12:19:25.101 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:25.101 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-0/keyring of=/dev/stdout 2026-04-17T12:19:25.171 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:25.171 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-1/keyring of=/dev/stdout 2026-04-17T12:19:25.237 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:25.237 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-2/keyring of=/dev/stdout 2026-04-17T12:19:25.306 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:25.307 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-3/keyring of=/dev/stdout 2026-04-17T12:19:25.375 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:25.375 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/mgr/ceph-x/keyring of=/dev/stdout 2026-04-17T12:19:25.402 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:25.402 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-4/keyring of=/dev/stdout 2026-04-17T12:19:25.470 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:25.470 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-5/keyring of=/dev/stdout 2026-04-17T12:19:25.537 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:25.538 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-6/keyring of=/dev/stdout 2026-04-17T12:19:25.606 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:25.606 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-7/keyring of=/dev/stdout 2026-04-17T12:19:25.673 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:25.673 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.client.0.keyring of=/dev/stdout 2026-04-17T12:19:25.692 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:25.692 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.client.1.keyring of=/dev/stdout 2026-04-17T12:19:25.732 DEBUG:teuthology.orchestra.run.vm08:> set -ex 2026-04-17T12:19:25.732 DEBUG:teuthology.orchestra.run.vm08:> dd if=/etc/ceph/ceph.client.2.keyring of=/dev/stdout 2026-04-17T12:19:25.751 INFO:tasks.ceph:Adding keys to all mons... 2026-04-17T12:19:25.751 DEBUG:teuthology.orchestra.run.vm02:> sudo tee -a /etc/ceph/ceph.keyring 2026-04-17T12:19:25.752 DEBUG:teuthology.orchestra.run.vm06:> sudo tee -a /etc/ceph/ceph.keyring 2026-04-17T12:19:25.774 INFO:teuthology.orchestra.run.vm02.stdout:[mgr.y] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBDJeJpYeuUJxAA/4VkBrgZnsWZ4/U//Mfo7w== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.0] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBFJeJpTHhpCRAA+6i0CevnCIRGZVTyWs6NcA== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.1] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBGJeJpaqWQABAAVJ7bi/3M4VKrdJHayI1RZA== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.2] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBGJeJpiQlgNRAAPToI/iErPmeRCUHb0EjUAg== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.3] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBHJeJp4EQgLBAA1lNGhH1i1YIGP2Mlhy6f6w== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[mgr.x] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBDJeJpU2+NKxAAyWAlidvCwr8a4Pp/M6YpLw== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.4] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBJJeJppmT3MBAA6a4rNlAjQDqRraKHOKSJ1g== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.5] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBKJeJp/bhdIhAA0YjtDncXnpxYA5lDwb04eg== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.6] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBLJeJptPEXIBAA/QLqLneq49k8SVfjgPNYKQ== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[osd.7] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBMJeJpeJsOFBAAPezmU/YWY4viUBAdarVObw== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[client.0] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBDJeJpRjhILhAAS3DlMXjDeEXIjsTaGLuRQA== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[client.1] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBDJeJp4CDRMRAA1eMnKkU/VaVXdj/Lrbn0cw== 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout:[client.2] 2026-04-17T12:19:25.775 INFO:teuthology.orchestra.run.vm02.stdout: key = AQBDJeJpKWuHNRAAjUeqMdM/hVGCcs+Di1SNKw== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[mgr.y] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBDJeJpYeuUJxAA/4VkBrgZnsWZ4/U//Mfo7w== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.0] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBFJeJpTHhpCRAA+6i0CevnCIRGZVTyWs6NcA== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.1] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBGJeJpaqWQABAAVJ7bi/3M4VKrdJHayI1RZA== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.2] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBGJeJpiQlgNRAAPToI/iErPmeRCUHb0EjUAg== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.3] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBHJeJp4EQgLBAA1lNGhH1i1YIGP2Mlhy6f6w== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[mgr.x] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBDJeJpU2+NKxAAyWAlidvCwr8a4Pp/M6YpLw== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.4] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBJJeJppmT3MBAA6a4rNlAjQDqRraKHOKSJ1g== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.5] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBKJeJp/bhdIhAA0YjtDncXnpxYA5lDwb04eg== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.6] 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBLJeJptPEXIBAA/QLqLneq49k8SVfjgPNYKQ== 2026-04-17T12:19:25.796 INFO:teuthology.orchestra.run.vm06.stdout:[osd.7] 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBMJeJpeJsOFBAAPezmU/YWY4viUBAdarVObw== 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBDJeJpRjhILhAAS3DlMXjDeEXIjsTaGLuRQA== 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout:[client.1] 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBDJeJp4CDRMRAA1eMnKkU/VaVXdj/Lrbn0cw== 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout:[client.2] 2026-04-17T12:19:25.797 INFO:teuthology.orchestra.run.vm06.stdout: key = AQBDJeJpKWuHNRAAjUeqMdM/hVGCcs+Di1SNKw== 2026-04-17T12:19:25.797 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.y --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-17T12:19:25.817 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.y --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-17T12:19:25.885 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.0 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:25.886 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.0 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:25.934 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.1 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:25.935 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.1 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:25.982 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.2 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:25.984 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.2 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.032 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.3 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.033 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.3 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.081 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.x --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-17T12:19:26.082 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=mgr.x --cap mon 'allow profile mgr' --cap osd 'allow *' --cap mds 'allow *' 2026-04-17T12:19:26.131 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.4 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.132 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.4 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.180 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.5 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.181 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.5 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.229 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.6 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.230 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.6 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.283 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.7 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.285 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=osd.7 --cap mon 'allow profile osd' --cap mgr 'allow profile osd' --cap osd 'allow *' 2026-04-17T12:19:26.333 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.0 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-17T12:19:26.335 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.0 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-17T12:19:26.384 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.1 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-17T12:19:26.385 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.1 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-17T12:19:26.448 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.2 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-17T12:19:26.449 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-authtool /etc/ceph/ceph.keyring --name=client.2 --cap mon 'allow rw' --cap mgr 'allow r' --cap osd 'allow rwx' --cap mds allow 2026-04-17T12:19:26.518 INFO:tasks.ceph:Running mkfs on mon nodes... 2026-04-17T12:19:26.518 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/mon/ceph-a 2026-04-17T12:19:26.549 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster ceph --mkfs -i a --monmap /home/ubuntu/cephtest/ceph.monmap --keyring /etc/ceph/ceph.keyring 2026-04-17T12:19:26.645 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/mon/ceph-a 2026-04-17T12:19:26.675 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /var/lib/ceph/mon/ceph-c 2026-04-17T12:19:26.747 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster ceph --mkfs -i c --monmap /home/ubuntu/cephtest/ceph.monmap --keyring /etc/ceph/ceph.keyring 2026-04-17T12:19:27.329 DEBUG:teuthology.orchestra.run.vm02:> sudo chown -R ceph:ceph /var/lib/ceph/mon/ceph-c 2026-04-17T12:19:27.357 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /var/lib/ceph/mon/ceph-b 2026-04-17T12:19:27.384 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph-mon --cluster ceph --mkfs -i b --monmap /home/ubuntu/cephtest/ceph.monmap --keyring /etc/ceph/ceph.keyring 2026-04-17T12:19:27.501 DEBUG:teuthology.orchestra.run.vm06:> sudo chown -R ceph:ceph /var/lib/ceph/mon/ceph-b 2026-04-17T12:19:27.529 DEBUG:teuthology.orchestra.run.vm02:> rm -- /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:27.531 DEBUG:teuthology.orchestra.run.vm06:> rm -- /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:19:27.588 INFO:tasks.ceph:Starting mon daemons in cluster ceph... 2026-04-17T12:19:27.589 INFO:tasks.ceph.mon.a:Restarting daemon 2026-04-17T12:19:27.589 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i a 2026-04-17T12:19:27.591 INFO:tasks.ceph.mon.a:Started 2026-04-17T12:19:27.591 INFO:tasks.ceph.mon.c:Restarting daemon 2026-04-17T12:19:27.592 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i c 2026-04-17T12:19:27.593 INFO:tasks.ceph.mon.c:Started 2026-04-17T12:19:27.593 INFO:tasks.ceph.mon.b:Restarting daemon 2026-04-17T12:19:27.593 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i b 2026-04-17T12:19:27.631 INFO:tasks.ceph.mon.b:Started 2026-04-17T12:19:27.631 INFO:tasks.ceph:Starting mgr daemons in cluster ceph... 2026-04-17T12:19:27.631 INFO:tasks.ceph.mgr.y:Restarting daemon 2026-04-17T12:19:27.631 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster ceph -i y 2026-04-17T12:19:27.633 INFO:tasks.ceph.mgr.y:Started 2026-04-17T12:19:27.633 INFO:tasks.ceph.mgr.x:Restarting daemon 2026-04-17T12:19:27.634 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster ceph -i x 2026-04-17T12:19:27.635 INFO:tasks.ceph.mgr.x:Started 2026-04-17T12:19:27.635 DEBUG:tasks.ceph:set 0 configs 2026-04-17T12:19:27.635 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph --cluster ceph config dump 2026-04-17T12:19:27.959 INFO:teuthology.orchestra.run.vm02.stdout:WHO MASK LEVEL OPTION VALUE RO 2026-04-17T12:19:27.970 INFO:tasks.ceph:Setting crush tunables to default 2026-04-17T12:19:27.970 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph --cluster ceph osd crush tunables default 2026-04-17T12:19:28.097 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-04-17T12:19:28.110 INFO:tasks.ceph:check_enable_crimson: False 2026-04-17T12:19:28.110 INFO:tasks.ceph:Starting osd daemons in cluster ceph... 2026-04-17T12:19:28.110 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:28.110 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-0/fsid of=/dev/stdout 2026-04-17T12:19:28.136 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:28.136 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-1/fsid of=/dev/stdout 2026-04-17T12:19:28.206 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:28.206 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-2/fsid of=/dev/stdout 2026-04-17T12:19:28.274 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:19:28.274 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/osd/ceph-3/fsid of=/dev/stdout 2026-04-17T12:19:28.337 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:28.337 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-4/fsid of=/dev/stdout 2026-04-17T12:19:28.364 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:28.364 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-5/fsid of=/dev/stdout 2026-04-17T12:19:28.431 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:28.431 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-6/fsid of=/dev/stdout 2026-04-17T12:19:28.501 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-04-17T12:19:28.501 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/osd/ceph-7/fsid of=/dev/stdout 2026-04-17T12:19:28.570 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new bcc8939e-938b-422b-8ddf-3dce21ad344b 0 2026-04-17T12:19:28.660 INFO:tasks.ceph.mgr.x.vm06.stderr:/usr/lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-04-17T12:19:28.660 INFO:tasks.ceph.mgr.x.vm06.stderr:Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-04-17T12:19:28.660 INFO:tasks.ceph.mgr.x.vm06.stderr: from numpy import show_config as show_numpy_config 2026-04-17T12:19:28.669 INFO:tasks.ceph.mgr.y.vm02.stderr:/usr/lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-04-17T12:19:28.669 INFO:tasks.ceph.mgr.y.vm02.stderr:Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-04-17T12:19:28.669 INFO:tasks.ceph.mgr.y.vm02.stderr: from numpy import show_config as show_numpy_config 2026-04-17T12:19:28.734 INFO:teuthology.orchestra.run.vm06.stdout:0 2026-04-17T12:19:28.743 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new 50a2a469-0b4d-4b52-aa07-46857fb19d3f 1 2026-04-17T12:19:28.867 INFO:teuthology.orchestra.run.vm06.stdout:1 2026-04-17T12:19:28.878 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new 4f83a37c-b579-4ce9-b3cc-db75ff9959f5 2 2026-04-17T12:19:29.006 INFO:teuthology.orchestra.run.vm06.stdout:2 2026-04-17T12:19:29.017 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new 379c5134-6a29-4bcb-aee4-38b693473064 3 2026-04-17T12:19:29.145 INFO:teuthology.orchestra.run.vm06.stdout:3 2026-04-17T12:19:29.156 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new ad85efa0-0f9d-4fe1-997a-8e936d8d38b7 4 2026-04-17T12:19:29.279 INFO:teuthology.orchestra.run.vm06.stdout:4 2026-04-17T12:19:29.290 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new ccf043d5-f57b-4519-84e4-b837a46c10db 5 2026-04-17T12:19:29.415 INFO:teuthology.orchestra.run.vm06.stdout:5 2026-04-17T12:19:29.426 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new b6201964-ccd3-417c-892f-572b94c96e8d 6 2026-04-17T12:19:29.548 INFO:teuthology.orchestra.run.vm06.stdout:6 2026-04-17T12:19:29.558 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph --cluster ceph osd new e81dca8b-137e-4b16-ad7c-5eab7af39b3a 7 2026-04-17T12:19:29.693 INFO:teuthology.orchestra.run.vm06.stdout:7 2026-04-17T12:19:29.704 INFO:tasks.ceph.osd.0:Restarting daemon 2026-04-17T12:19:29.704 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 0 2026-04-17T12:19:29.706 INFO:tasks.ceph.osd.0:Started 2026-04-17T12:19:29.706 INFO:tasks.ceph.osd.1:Restarting daemon 2026-04-17T12:19:29.706 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 1 2026-04-17T12:19:29.707 INFO:tasks.ceph.osd.1:Started 2026-04-17T12:19:29.707 INFO:tasks.ceph.osd.2:Restarting daemon 2026-04-17T12:19:29.707 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 2 2026-04-17T12:19:29.709 INFO:tasks.ceph.osd.2:Started 2026-04-17T12:19:29.709 INFO:tasks.ceph.osd.3:Restarting daemon 2026-04-17T12:19:29.709 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 3 2026-04-17T12:19:29.711 INFO:tasks.ceph.osd.3:Started 2026-04-17T12:19:29.711 INFO:tasks.ceph.osd.4:Restarting daemon 2026-04-17T12:19:29.711 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 4 2026-04-17T12:19:29.713 INFO:tasks.ceph.osd.4:Started 2026-04-17T12:19:29.713 INFO:tasks.ceph.osd.5:Restarting daemon 2026-04-17T12:19:29.713 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 5 2026-04-17T12:19:29.714 INFO:tasks.ceph.osd.5:Started 2026-04-17T12:19:29.714 INFO:tasks.ceph.osd.6:Restarting daemon 2026-04-17T12:19:29.714 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 6 2026-04-17T12:19:29.716 INFO:tasks.ceph.osd.6:Started 2026-04-17T12:19:29.716 INFO:tasks.ceph.osd.7:Restarting daemon 2026-04-17T12:19:29.716 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-osd -f --cluster ceph -i 7 2026-04-17T12:19:29.723 INFO:tasks.ceph.osd.7:Started 2026-04-17T12:19:29.723 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json 2026-04-17T12:19:29.890 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:29.890 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":10,"fsid":"50e117d9-5cfb-424b-9f73-1b1dada394f8","created":"2026-04-17T12:19:27.890556+0000","modified":"2026-04-17T12:19:29.687237+0000","last_up_change":"0.000000","last_in_change":"2026-04-17T12:19:29.687237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"bcc8939e-938b-422b-8ddf-3dce21ad344b","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"50a2a469-0b4d-4b52-aa07-46857fb19d3f","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"4f83a37c-b579-4ce9-b3cc-db75ff9959f5","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"379c5134-6a29-4bcb-aee4-38b693473064","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"ad85efa0-0f9d-4fe1-997a-8e936d8d38b7","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"ccf043d5-f57b-4519-84e4-b837a46c10db","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":6,"uuid":"b6201964-ccd3-417c-892f-572b94c96e8d","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":7,"uuid":"e81dca8b-137e-4b16-ad7c-5eab7af39b3a","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-17T12:19:29.892 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:19:29.891+0000 7f518921b900 -1 Falling back to public interface 2026-04-17T12:19:29.899 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:19:29.898+0000 7fe104387900 -1 Falling back to public interface 2026-04-17T12:19:29.901 INFO:tasks.ceph.ceph_manager.ceph:[] 2026-04-17T12:19:29.901 INFO:tasks.ceph:Waiting for OSDs to come up 2026-04-17T12:19:29.905 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:19:29.904+0000 7f903b93c900 -1 Falling back to public interface 2026-04-17T12:19:29.906 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:19:29.905+0000 7f5842812900 -1 Falling back to public interface 2026-04-17T12:19:29.912 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:19:29.911+0000 7f729ea27900 -1 Falling back to public interface 2026-04-17T12:19:29.917 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:19:29.915+0000 7fd89a341900 -1 Falling back to public interface 2026-04-17T12:19:29.917 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:19:29.915+0000 7f2182a0f900 -1 Falling back to public interface 2026-04-17T12:19:29.921 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:19:29.920+0000 7f9eee521900 -1 Falling back to public interface 2026-04-17T12:19:30.322 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:19:30.319+0000 7f518921b900 -1 osd.2 0 log_to_monitors true 2026-04-17T12:19:30.361 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:19:30.360+0000 7f903b93c900 -1 osd.4 0 log_to_monitors true 2026-04-17T12:19:30.367 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:19:30.366+0000 7f5842812900 -1 osd.1 0 log_to_monitors true 2026-04-17T12:19:30.373 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:19:30.373+0000 7fe104387900 -1 osd.0 0 log_to_monitors true 2026-04-17T12:19:30.381 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:19:30.380+0000 7f9eee521900 -1 osd.5 0 log_to_monitors true 2026-04-17T12:19:30.409 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:19:30.408+0000 7f729ea27900 -1 osd.6 0 log_to_monitors true 2026-04-17T12:19:30.438 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:19:30.437+0000 7fd89a341900 -1 osd.7 0 log_to_monitors true 2026-04-17T12:19:30.475 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:19:30.474+0000 7f2182a0f900 -1 osd.3 0 log_to_monitors true 2026-04-17T12:19:30.707 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster ceph osd dump --format=json 2026-04-17T12:19:30.821 INFO:teuthology.misc.health.vm02.stdout: 2026-04-17T12:19:30.821 INFO:teuthology.misc.health.vm02.stdout:{"epoch":10,"fsid":"50e117d9-5cfb-424b-9f73-1b1dada394f8","created":"2026-04-17T12:19:27.890556+0000","modified":"2026-04-17T12:19:29.687237+0000","last_up_change":"0.000000","last_in_change":"2026-04-17T12:19:29.687237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":2,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"bcc8939e-938b-422b-8ddf-3dce21ad344b","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":1,"uuid":"50a2a469-0b4d-4b52-aa07-46857fb19d3f","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":2,"uuid":"4f83a37c-b579-4ce9-b3cc-db75ff9959f5","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":3,"uuid":"379c5134-6a29-4bcb-aee4-38b693473064","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":4,"uuid":"ad85efa0-0f9d-4fe1-997a-8e936d8d38b7","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":5,"uuid":"ccf043d5-f57b-4519-84e4-b837a46c10db","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":6,"uuid":"b6201964-ccd3-417c-892f-572b94c96e8d","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]},{"osd":7,"uuid":"e81dca8b-137e-4b16-ad7c-5eab7af39b3a","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":0,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[]},"cluster_addrs":{"addrvec":[]},"heartbeat_back_addrs":{"addrvec":[]},"heartbeat_front_addrs":{"addrvec":[]},"public_addr":"(unrecognized address family 0)/0","cluster_addr":"(unrecognized address family 0)/0","heartbeat_back_addr":"(unrecognized address family 0)/0","heartbeat_front_addr":"(unrecognized address family 0)/0","state":["exists","new"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":0,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-17T12:19:30.830 DEBUG:teuthology.misc:0 of 8 OSDs are up 2026-04-17T12:19:31.558 INFO:tasks.ceph.mgr.x.vm06.stderr:2026-04-17T12:19:31.557+0000 7f801d0f1640 -1 mgr.server handle_report got status from non-daemon mon.b 2026-04-17T12:19:31.559 INFO:tasks.ceph.mgr.x.vm06.stderr:2026-04-17T12:19:31.558+0000 7f801d0f1640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-04-17T12:19:35.893 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:19:35.892+0000 7f729a1d1640 -1 osd.6 0 waiting for initial osdmap 2026-04-17T12:19:35.906 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:19:35.904+0000 7fd896ae1640 -1 osd.7 0 waiting for initial osdmap 2026-04-17T12:19:35.912 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:19:35.910+0000 7f7294fd5640 -1 osd.6 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:35.928 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:19:35.927+0000 7fd8910d3640 -1 osd.7 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:36.210 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:19:36.209+0000 7f90378ca640 -1 osd.4 0 waiting for initial osdmap 2026-04-17T12:19:36.228 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:19:36.227+0000 7f90326ce640 -1 osd.4 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:36.233 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:19:36.232+0000 7f9eea4af640 -1 osd.5 0 waiting for initial osdmap 2026-04-17T12:19:36.250 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:19:36.249+0000 7f9ee52b3640 -1 osd.5 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:36.444 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:19:36.443+0000 7f51849a9640 -1 osd.2 0 waiting for initial osdmap 2026-04-17T12:19:36.460 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:19:36.458+0000 7f517f7ad640 -1 osd.2 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:36.510 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:19:36.508+0000 7f217e175640 -1 osd.3 0 waiting for initial osdmap 2026-04-17T12:19:36.510 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:19:36.509+0000 7fe100b27640 -1 osd.0 0 waiting for initial osdmap 2026-04-17T12:19:36.525 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:19:36.524+0000 7f583df89640 -1 osd.1 0 waiting for initial osdmap 2026-04-17T12:19:36.531 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:19:36.530+0000 7fe0fb119640 -1 osd.0 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:36.531 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:19:36.530+0000 7f2178f79640 -1 osd.3 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:36.537 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:19:36.536+0000 7f5838d8d640 -1 osd.1 12 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-04-17T12:19:37.634 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage ceph --cluster ceph osd dump --format=json 2026-04-17T12:19:37.828 INFO:teuthology.misc.health.vm02.stdout: 2026-04-17T12:19:37.828 INFO:teuthology.misc.health.vm02.stdout:{"epoch":13,"fsid":"50e117d9-5cfb-424b-9f73-1b1dada394f8","created":"2026-04-17T12:19:27.890556+0000","modified":"2026-04-17T12:19:36.908279+0000","last_up_change":"2026-04-17T12:19:36.908279+0000","last_in_change":"2026-04-17T12:19:29.687237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":0,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[],"osds":[{"osd":0,"uuid":"bcc8939e-938b-422b-8ddf-3dce21ad344b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6809","nonce":812331292}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6811","nonce":812331292}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6815","nonce":812331292}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6813","nonce":812331292}]},"public_addr":"192.168.123.102:6809/812331292","cluster_addr":"192.168.123.102:6811/812331292","heartbeat_back_addr":"192.168.123.102:6815/812331292","heartbeat_front_addr":"192.168.123.102:6813/812331292","state":["exists","up"]},{"osd":1,"uuid":"50a2a469-0b4d-4b52-aa07-46857fb19d3f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6817","nonce":1730862675}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6819","nonce":1730862675}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6823","nonce":1730862675}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6821","nonce":1730862675}]},"public_addr":"192.168.123.102:6817/1730862675","cluster_addr":"192.168.123.102:6819/1730862675","heartbeat_back_addr":"192.168.123.102:6823/1730862675","heartbeat_front_addr":"192.168.123.102:6821/1730862675","state":["exists","up"]},{"osd":2,"uuid":"4f83a37c-b579-4ce9-b3cc-db75ff9959f5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6801","nonce":4174437851}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6803","nonce":4174437851}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6807","nonce":4174437851}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6805","nonce":4174437851}]},"public_addr":"192.168.123.102:6801/4174437851","cluster_addr":"192.168.123.102:6803/4174437851","heartbeat_back_addr":"192.168.123.102:6807/4174437851","heartbeat_front_addr":"192.168.123.102:6805/4174437851","state":["exists","up"]},{"osd":3,"uuid":"379c5134-6a29-4bcb-aee4-38b693473064","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6824","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6825","nonce":1644882209}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6826","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6827","nonce":1644882209}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6830","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6831","nonce":1644882209}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6828","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6829","nonce":1644882209}]},"public_addr":"192.168.123.102:6825/1644882209","cluster_addr":"192.168.123.102:6827/1644882209","heartbeat_back_addr":"192.168.123.102:6831/1644882209","heartbeat_front_addr":"192.168.123.102:6829/1644882209","state":["exists","up"]},{"osd":4,"uuid":"ad85efa0-0f9d-4fe1-997a-8e936d8d38b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6801","nonce":3059241276}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6803","nonce":3059241276}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6807","nonce":3059241276}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6805","nonce":3059241276}]},"public_addr":"192.168.123.106:6801/3059241276","cluster_addr":"192.168.123.106:6803/3059241276","heartbeat_back_addr":"192.168.123.106:6807/3059241276","heartbeat_front_addr":"192.168.123.106:6805/3059241276","state":["exists","up"]},{"osd":5,"uuid":"ccf043d5-f57b-4519-84e4-b837a46c10db","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6825","nonce":752206696}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6827","nonce":752206696}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6831","nonce":752206696}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6829","nonce":752206696}]},"public_addr":"192.168.123.106:6825/752206696","cluster_addr":"192.168.123.106:6827/752206696","heartbeat_back_addr":"192.168.123.106:6831/752206696","heartbeat_front_addr":"192.168.123.106:6829/752206696","state":["exists","up"]},{"osd":6,"uuid":"b6201964-ccd3-417c-892f-572b94c96e8d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6809","nonce":74249396}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6811","nonce":74249396}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6815","nonce":74249396}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6813","nonce":74249396}]},"public_addr":"192.168.123.106:6809/74249396","cluster_addr":"192.168.123.106:6811/74249396","heartbeat_back_addr":"192.168.123.106:6815/74249396","heartbeat_front_addr":"192.168.123.106:6813/74249396","state":["exists","up"]},{"osd":7,"uuid":"e81dca8b-137e-4b16-ad7c-5eab7af39b3a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6817","nonce":4261841531}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6819","nonce":4261841531}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6823","nonce":4261841531}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6821","nonce":4261841531}]},"public_addr":"192.168.123.106:6817/4261841531","cluster_addr":"192.168.123.106:6819/4261841531","heartbeat_back_addr":"192.168.123.106:6823/4261841531","heartbeat_front_addr":"192.168.123.106:6821/4261841531","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-17T12:19:37.839 DEBUG:teuthology.misc:8 of 8 OSDs are up 2026-04-17T12:19:37.840 INFO:tasks.ceph:Creating RBD pool 2026-04-17T12:19:37.840 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph --cluster ceph osd pool create rbd 8 2026-04-17T12:19:38.962 INFO:teuthology.orchestra.run.vm02.stderr:pool 'rbd' created 2026-04-17T12:19:38.975 DEBUG:teuthology.orchestra.run.vm02:> rbd --cluster ceph pool init rbd 2026-04-17T12:19:39.013 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:19:39.013 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:19:42.018 INFO:tasks.ceph:Starting mds daemons in cluster ceph... 2026-04-17T12:19:42.018 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph config log 1 --format=json 2026-04-17T12:19:42.018 INFO:tasks.daemonwatchdog.daemon_watchdog:watchdog starting 2026-04-17T12:19:42.268 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:42.280 INFO:teuthology.orchestra.run.vm02.stdout:[{"version":9,"timestamp":"2026-04-17T12:19:36.525765+0000","name":"","changes":[{"name":"osd.1/osd_mclock_max_capacity_iops_hdd","new_value":"3289.621926"}]}] 2026-04-17T12:19:42.280 INFO:tasks.ceph_manager:config epoch is 9 2026-04-17T12:19:42.280 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-04-17T12:19:42.280 INFO:tasks.ceph.ceph_manager.ceph:waiting for mgr available 2026-04-17T12:19:42.280 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph mgr dump --format=json 2026-04-17T12:19:42.523 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:42.537 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":5,"flags":0,"active_gid":4108,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":836427380},{"type":"v1","addr":"192.168.123.106:6833","nonce":836427380}]},"active_addr":"192.168.123.106:6833/836427380","active_change":"2026-04-17T12:19:30.537383+0000","active_mgr_features":4541880224203014143,"available":true,"standbys":[{"gid":4111,"name":"y","mgr_features":4541880224203014143,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.2.0","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":true,"error_string":"","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["iostat","nfs"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.2.0","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":true,"error_string":"","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"tentacle":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":0,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":357544963}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":31400782}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":3397297258}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":2433296907}]}]} 2026-04-17T12:19:42.538 INFO:tasks.ceph.ceph_manager.ceph:mgr available! 2026-04-17T12:19:42.538 INFO:tasks.ceph.ceph_manager.ceph:waiting for all up 2026-04-17T12:19:42.538 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json 2026-04-17T12:19:42.749 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:42.749 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":18,"fsid":"50e117d9-5cfb-424b-9f73-1b1dada394f8","created":"2026-04-17T12:19:27.890556+0000","modified":"2026-04-17T12:19:42.004747+0000","last_up_change":"2026-04-17T12:19:36.908279+0000","last_in_change":"2026-04-17T12:19:29.687237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":"rbd","create_time":"2026-04-17T12:19:38.066334+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":18,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2,"score_stable":2,"optimal_score":1,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":".mgr","create_time":"2026-04-17T12:19:38.574835+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"17","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":8,"score_stable":8,"optimal_score":0.25,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"bcc8939e-938b-422b-8ddf-3dce21ad344b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6809","nonce":812331292}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6811","nonce":812331292}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6815","nonce":812331292}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6813","nonce":812331292}]},"public_addr":"192.168.123.102:6809/812331292","cluster_addr":"192.168.123.102:6811/812331292","heartbeat_back_addr":"192.168.123.102:6815/812331292","heartbeat_front_addr":"192.168.123.102:6813/812331292","state":["exists","up"]},{"osd":1,"uuid":"50a2a469-0b4d-4b52-aa07-46857fb19d3f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6817","nonce":1730862675}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6819","nonce":1730862675}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6823","nonce":1730862675}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6821","nonce":1730862675}]},"public_addr":"192.168.123.102:6817/1730862675","cluster_addr":"192.168.123.102:6819/1730862675","heartbeat_back_addr":"192.168.123.102:6823/1730862675","heartbeat_front_addr":"192.168.123.102:6821/1730862675","state":["exists","up"]},{"osd":2,"uuid":"4f83a37c-b579-4ce9-b3cc-db75ff9959f5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6801","nonce":4174437851}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6803","nonce":4174437851}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6807","nonce":4174437851}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6805","nonce":4174437851}]},"public_addr":"192.168.123.102:6801/4174437851","cluster_addr":"192.168.123.102:6803/4174437851","heartbeat_back_addr":"192.168.123.102:6807/4174437851","heartbeat_front_addr":"192.168.123.102:6805/4174437851","state":["exists","up"]},{"osd":3,"uuid":"379c5134-6a29-4bcb-aee4-38b693473064","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6824","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6825","nonce":1644882209}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6826","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6827","nonce":1644882209}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6830","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6831","nonce":1644882209}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6828","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6829","nonce":1644882209}]},"public_addr":"192.168.123.102:6825/1644882209","cluster_addr":"192.168.123.102:6827/1644882209","heartbeat_back_addr":"192.168.123.102:6831/1644882209","heartbeat_front_addr":"192.168.123.102:6829/1644882209","state":["exists","up"]},{"osd":4,"uuid":"ad85efa0-0f9d-4fe1-997a-8e936d8d38b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6801","nonce":3059241276}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6803","nonce":3059241276}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6807","nonce":3059241276}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6805","nonce":3059241276}]},"public_addr":"192.168.123.106:6801/3059241276","cluster_addr":"192.168.123.106:6803/3059241276","heartbeat_back_addr":"192.168.123.106:6807/3059241276","heartbeat_front_addr":"192.168.123.106:6805/3059241276","state":["exists","up"]},{"osd":5,"uuid":"ccf043d5-f57b-4519-84e4-b837a46c10db","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6825","nonce":752206696}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6827","nonce":752206696}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6831","nonce":752206696}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6829","nonce":752206696}]},"public_addr":"192.168.123.106:6825/752206696","cluster_addr":"192.168.123.106:6827/752206696","heartbeat_back_addr":"192.168.123.106:6831/752206696","heartbeat_front_addr":"192.168.123.106:6829/752206696","state":["exists","up"]},{"osd":6,"uuid":"b6201964-ccd3-417c-892f-572b94c96e8d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6809","nonce":74249396}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6811","nonce":74249396}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6815","nonce":74249396}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6813","nonce":74249396}]},"public_addr":"192.168.123.106:6809/74249396","cluster_addr":"192.168.123.106:6811/74249396","heartbeat_back_addr":"192.168.123.106:6815/74249396","heartbeat_front_addr":"192.168.123.106:6813/74249396","state":["exists","up"]},{"osd":7,"uuid":"e81dca8b-137e-4b16-ad7c-5eab7af39b3a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6817","nonce":4261841531}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6819","nonce":4261841531}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6823","nonce":4261841531}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6821","nonce":4261841531}]},"public_addr":"192.168.123.106:6817/4261841531","cluster_addr":"192.168.123.106:6819/4261841531","heartbeat_back_addr":"192.168.123.106:6823/4261841531","heartbeat_front_addr":"192.168.123.106:6821/4261841531","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.361276+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.361031+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.317897+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.478377+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-17T12:19:42.763 INFO:tasks.ceph.ceph_manager.ceph:all up! 2026-04-17T12:19:42.763 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd dump --format=json 2026-04-17T12:19:42.962 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:42.962 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":18,"fsid":"50e117d9-5cfb-424b-9f73-1b1dada394f8","created":"2026-04-17T12:19:27.890556+0000","modified":"2026-04-17T12:19:42.004747+0000","last_up_change":"2026-04-17T12:19:36.908279+0000","last_in_change":"2026-04-17T12:19:29.687237+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":4,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":"rbd","create_time":"2026-04-17T12:19:38.066334+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":8,"pg_placement_num":8,"pg_placement_num_target":8,"pg_num_target":8,"pg_num_pending":8,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"18","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":18,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":2,"score_stable":2,"optimal_score":1,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":".mgr","create_time":"2026-04-17T12:19:38.574835+0000","flags":1,"flags_names":"hashpspool","type":1,"size":2,"min_size":1,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"17","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":8,"score_stable":8,"optimal_score":0.25,"raw_score_acting":2,"raw_score_stable":2,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"bcc8939e-938b-422b-8ddf-3dce21ad344b","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6809","nonce":812331292}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6811","nonce":812331292}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6815","nonce":812331292}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":812331292},{"type":"v1","addr":"192.168.123.102:6813","nonce":812331292}]},"public_addr":"192.168.123.102:6809/812331292","cluster_addr":"192.168.123.102:6811/812331292","heartbeat_back_addr":"192.168.123.102:6815/812331292","heartbeat_front_addr":"192.168.123.102:6813/812331292","state":["exists","up"]},{"osd":1,"uuid":"50a2a469-0b4d-4b52-aa07-46857fb19d3f","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6817","nonce":1730862675}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6818","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6819","nonce":1730862675}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6822","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6823","nonce":1730862675}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6820","nonce":1730862675},{"type":"v1","addr":"192.168.123.102:6821","nonce":1730862675}]},"public_addr":"192.168.123.102:6817/1730862675","cluster_addr":"192.168.123.102:6819/1730862675","heartbeat_back_addr":"192.168.123.102:6823/1730862675","heartbeat_front_addr":"192.168.123.102:6821/1730862675","state":["exists","up"]},{"osd":2,"uuid":"4f83a37c-b579-4ce9-b3cc-db75ff9959f5","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6801","nonce":4174437851}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6803","nonce":4174437851}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6807","nonce":4174437851}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":4174437851},{"type":"v1","addr":"192.168.123.102:6805","nonce":4174437851}]},"public_addr":"192.168.123.102:6801/4174437851","cluster_addr":"192.168.123.102:6803/4174437851","heartbeat_back_addr":"192.168.123.102:6807/4174437851","heartbeat_front_addr":"192.168.123.102:6805/4174437851","state":["exists","up"]},{"osd":3,"uuid":"379c5134-6a29-4bcb-aee4-38b693473064","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6824","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6825","nonce":1644882209}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6826","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6827","nonce":1644882209}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6830","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6831","nonce":1644882209}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6828","nonce":1644882209},{"type":"v1","addr":"192.168.123.102:6829","nonce":1644882209}]},"public_addr":"192.168.123.102:6825/1644882209","cluster_addr":"192.168.123.102:6827/1644882209","heartbeat_back_addr":"192.168.123.102:6831/1644882209","heartbeat_front_addr":"192.168.123.102:6829/1644882209","state":["exists","up"]},{"osd":4,"uuid":"ad85efa0-0f9d-4fe1-997a-8e936d8d38b7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6801","nonce":3059241276}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6803","nonce":3059241276}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6807","nonce":3059241276}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":3059241276},{"type":"v1","addr":"192.168.123.106:6805","nonce":3059241276}]},"public_addr":"192.168.123.106:6801/3059241276","cluster_addr":"192.168.123.106:6803/3059241276","heartbeat_back_addr":"192.168.123.106:6807/3059241276","heartbeat_front_addr":"192.168.123.106:6805/3059241276","state":["exists","up"]},{"osd":5,"uuid":"ccf043d5-f57b-4519-84e4-b837a46c10db","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6825","nonce":752206696}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6827","nonce":752206696}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6831","nonce":752206696}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":752206696},{"type":"v1","addr":"192.168.123.106:6829","nonce":752206696}]},"public_addr":"192.168.123.106:6825/752206696","cluster_addr":"192.168.123.106:6827/752206696","heartbeat_back_addr":"192.168.123.106:6831/752206696","heartbeat_front_addr":"192.168.123.106:6829/752206696","state":["exists","up"]},{"osd":6,"uuid":"b6201964-ccd3-417c-892f-572b94c96e8d","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6809","nonce":74249396}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6811","nonce":74249396}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6815","nonce":74249396}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":74249396},{"type":"v1","addr":"192.168.123.106:6813","nonce":74249396}]},"public_addr":"192.168.123.106:6809/74249396","cluster_addr":"192.168.123.106:6811/74249396","heartbeat_back_addr":"192.168.123.106:6815/74249396","heartbeat_front_addr":"192.168.123.106:6813/74249396","state":["exists","up"]},{"osd":7,"uuid":"e81dca8b-137e-4b16-ad7c-5eab7af39b3a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":13,"up_thru":15,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6817","nonce":4261841531}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6819","nonce":4261841531}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6823","nonce":4261841531}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":4261841531},{"type":"v1","addr":"192.168.123.106:6821","nonce":4261841531}]},"public_addr":"192.168.123.106:6817/4261841531","cluster_addr":"192.168.123.106:6819/4261841531","heartbeat_back_addr":"192.168.123.106:6823/4261841531","heartbeat_front_addr":"192.168.123.106:6821/4261841531","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.361276+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.361031+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.317897+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"2026-04-17T12:19:31.478377+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4541880224203014143,"old_weight":0,"last_purged_snaps_scrub":"0.000000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[{"pool":1,"snaps":[{"begin":2,"length":1}]}],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-04-17T12:19:42.975 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.0 flush_pg_stats 2026-04-17T12:19:42.975 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.1 flush_pg_stats 2026-04-17T12:19:42.975 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.2 flush_pg_stats 2026-04-17T12:19:42.976 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.3 flush_pg_stats 2026-04-17T12:19:42.976 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.4 flush_pg_stats 2026-04-17T12:19:42.976 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.5 flush_pg_stats 2026-04-17T12:19:42.976 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.6 flush_pg_stats 2026-04-17T12:19:42.976 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph tell osd.7 flush_pg_stats 2026-04-17T12:19:43.183 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.183 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.4 2026-04-17T12:19:43.190 INFO:teuthology.orchestra.run.vm02.stdout:55834574852 2026-04-17T12:19:43.190 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.1 2026-04-17T12:19:43.194 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.195 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.5 2026-04-17T12:19:43.196 INFO:teuthology.orchestra.run.vm02.stdout:55834574852 2026-04-17T12:19:43.197 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.2 2026-04-17T12:19:43.206 INFO:teuthology.orchestra.run.vm02.stdout:55834574852 2026-04-17T12:19:43.206 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.0 2026-04-17T12:19:43.207 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.207 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.3 2026-04-17T12:19:43.210 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.211 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.7 2026-04-17T12:19:43.226 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.226 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.6 2026-04-17T12:19:43.535 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.547 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.551 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574851 for osd.1 2026-04-17T12:19:43.571 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574851 for osd.0 2026-04-17T12:19:43.575 INFO:teuthology.orchestra.run.vm02.stdout:55834574850 2026-04-17T12:19:43.621 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574850 for osd.4 2026-04-17T12:19:43.646 INFO:teuthology.orchestra.run.vm02.stdout:55834574850 2026-04-17T12:19:43.652 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:43.655 INFO:teuthology.orchestra.run.vm02.stdout:55834574850 2026-04-17T12:19:43.664 INFO:teuthology.orchestra.run.vm02.stdout:55834574850 2026-04-17T12:19:43.668 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574850 for osd.5 2026-04-17T12:19:43.669 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574850 for osd.7 2026-04-17T12:19:43.677 INFO:teuthology.orchestra.run.vm02.stdout:55834574850 2026-04-17T12:19:43.678 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574851 for osd.2 2026-04-17T12:19:43.686 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574850 for osd.3 2026-04-17T12:19:43.690 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574850 for osd.6 2026-04-17T12:19:44.553 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.1 2026-04-17T12:19:44.572 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.0 2026-04-17T12:19:44.622 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.4 2026-04-17T12:19:44.668 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.5 2026-04-17T12:19:44.670 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.7 2026-04-17T12:19:44.679 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.2 2026-04-17T12:19:44.687 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.3 2026-04-17T12:19:44.691 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.6 2026-04-17T12:19:44.856 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:44.885 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574851 for osd.1 2026-04-17T12:19:44.930 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:44.948 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574851 for osd.0 2026-04-17T12:19:44.972 INFO:teuthology.orchestra.run.vm02.stdout:55834574850 2026-04-17T12:19:44.991 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:44.992 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574850 for osd.4 2026-04-17T12:19:45.025 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574851 for osd.2 2026-04-17T12:19:45.030 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:45.049 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.7 2026-04-17T12:19:45.049 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:45.065 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:45.080 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.5 2026-04-17T12:19:45.080 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:45.089 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:45.101 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.6 2026-04-17T12:19:45.101 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:45.104 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:45.116 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.3 2026-04-17T12:19:45.116 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:45.885 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.1 2026-04-17T12:19:45.949 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.0 2026-04-17T12:19:45.993 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.4 2026-04-17T12:19:46.026 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph osd last-stat-seq osd.2 2026-04-17T12:19:46.111 INFO:teuthology.orchestra.run.vm02.stdout:55834574852 2026-04-17T12:19:46.130 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.1 2026-04-17T12:19:46.130 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:46.168 INFO:teuthology.orchestra.run.vm02.stdout:55834574852 2026-04-17T12:19:46.186 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.0 2026-04-17T12:19:46.186 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:46.217 INFO:teuthology.orchestra.run.vm02.stdout:55834574851 2026-04-17T12:19:46.229 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574851 got 55834574851 for osd.4 2026-04-17T12:19:46.229 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:46.258 INFO:teuthology.orchestra.run.vm02.stdout:55834574852 2026-04-17T12:19:46.271 INFO:tasks.ceph.ceph_manager.ceph:need seq 55834574852 got 55834574852 for osd.2 2026-04-17T12:19:46.271 DEBUG:teuthology.parallel:result is None 2026-04-17T12:19:46.271 INFO:tasks.ceph.ceph_manager.ceph:waiting for clean 2026-04-17T12:19:46.271 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json 2026-04-17T12:19:46.515 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:46.515 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-04-17T12:19:46.528 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":15,"stamp":"2026-04-17T12:19:44.550201+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":11,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":8,"kb":754974720,"kb_used":217300,"kb_used_data":2260,"kb_used_omap":50,"kb_used_meta":214477,"kb_avail":754757420,"statfs":{"total":773094113280,"available":772871598080,"internally_reserved":0,"allocated":2314240,"data_stored":1352828,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":51982,"internal_metadata":219624690},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":11,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"2.541754"},"pg_stats":[{"pgid":"1.7","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.713442+0000","last_change":"2026-04-17T12:19:42.713568+0000","last_active":"2026-04-17T12:19:42.713442+0000","last_peered":"2026-04-17T12:19:42.713442+0000","last_clean":"2026-04-17T12:19:42.713442+0000","last_became_active":"2026-04-17T12:19:40.013449+0000","last_became_peered":"2026-04-17T12:19:40.013449+0000","last_unstale":"2026-04-17T12:19:42.713442+0000","last_undegraded":"2026-04-17T12:19:42.713442+0000","last_fullsized":"2026-04-17T12:19:42.713442+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T15:47:10.055969+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000232876,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,3],"acting":[6,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.6","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.713419+0000","last_change":"2026-04-17T12:19:42.713542+0000","last_active":"2026-04-17T12:19:42.713419+0000","last_peered":"2026-04-17T12:19:42.713419+0000","last_clean":"2026-04-17T12:19:42.713419+0000","last_became_active":"2026-04-17T12:19:40.015415+0000","last_became_peered":"2026-04-17T12:19:40.015415+0000","last_unstale":"2026-04-17T12:19:42.713419+0000","last_undegraded":"2026-04-17T12:19:42.713419+0000","last_fullsized":"2026-04-17T12:19:42.713419+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T19:23:21.349991+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000284181,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0],"acting":[6,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.5","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.009970+0000","last_change":"2026-04-17T12:19:42.010374+0000","last_active":"2026-04-17T12:19:42.009970+0000","last_peered":"2026-04-17T12:19:42.009970+0000","last_clean":"2026-04-17T12:19:42.009970+0000","last_became_active":"2026-04-17T12:19:40.329751+0000","last_became_peered":"2026-04-17T12:19:40.329751+0000","last_unstale":"2026-04-17T12:19:42.009970+0000","last_undegraded":"2026-04-17T12:19:42.009970+0000","last_fullsized":"2026-04-17T12:19:42.009970+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:54:57.330537+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00051761799999999996,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2],"acting":[4,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.4","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.181017+0000","last_change":"2026-04-17T12:19:42.181110+0000","last_active":"2026-04-17T12:19:42.181017+0000","last_peered":"2026-04-17T12:19:42.181017+0000","last_clean":"2026-04-17T12:19:42.181017+0000","last_became_active":"2026-04-17T12:19:40.013650+0000","last_became_peered":"2026-04-17T12:19:40.013650+0000","last_unstale":"2026-04-17T12:19:42.181017+0000","last_undegraded":"2026-04-17T12:19:42.181017+0000","last_fullsized":"2026-04-17T12:19:42.181017+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T19:23:21.349991+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000293219,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5],"acting":[3,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"16'32","reported_seq":59,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.012321+0000","last_change":"2026-04-17T12:19:40.013859+0000","last_active":"2026-04-17T12:19:42.012321+0000","last_peered":"2026-04-17T12:19:42.012321+0000","last_clean":"2026-04-17T12:19:42.012321+0000","last_became_active":"2026-04-17T12:19:40.013721+0000","last_became_peered":"2026-04-17T12:19:40.013721+0000","last_unstale":"2026-04-17T12:19:42.012321+0000","last_undegraded":"2026-04-17T12:19:42.012321+0000","last_fullsized":"2026-04-17T12:19:42.012321+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T13:22:59.361909+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,1],"acting":[7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"1.3","version":"16'1","reported_seq":14,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-17T12:19:41.009013+0000","last_change":"2026-04-17T12:19:40.014326+0000","last_active":"2026-04-17T12:19:41.009013+0000","last_peered":"2026-04-17T12:19:41.009013+0000","last_clean":"2026-04-17T12:19:41.009013+0000","last_became_active":"2026-04-17T12:19:40.013706+0000","last_became_peered":"2026-04-17T12:19:40.013706+0000","last_unstale":"2026-04-17T12:19:41.009013+0000","last_undegraded":"2026-04-17T12:19:41.009013+0000","last_fullsized":"2026-04-17T12:19:41.009013+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:54:57.330537+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6],"acting":[1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"1.2","version":"18'2","reported_seq":22,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.013052+0000","last_change":"2026-04-17T12:19:42.013052+0000","last_active":"2026-04-17T12:19:42.013052+0000","last_peered":"2026-04-17T12:19:42.013052+0000","last_clean":"2026-04-17T12:19:42.013052+0000","last_became_active":"2026-04-17T12:19:40.015414+0000","last_became_peered":"2026-04-17T12:19:40.015414+0000","last_unstale":"2026-04-17T12:19:42.013052+0000","last_undegraded":"2026-04-17T12:19:42.013052+0000","last_fullsized":"2026-04-17T12:19:42.013052+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T13:03:17.584995+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00055161299999999995,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.1","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.010176+0000","last_change":"2026-04-17T12:19:42.010320+0000","last_active":"2026-04-17T12:19:42.010176+0000","last_peered":"2026-04-17T12:19:42.010176+0000","last_clean":"2026-04-17T12:19:42.010176+0000","last_became_active":"2026-04-17T12:19:40.015608+0000","last_became_peered":"2026-04-17T12:19:40.015608+0000","last_unstale":"2026-04-17T12:19:42.010176+0000","last_undegraded":"2026-04-17T12:19:42.010176+0000","last_fullsized":"2026-04-17T12:19:42.010176+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:54:57.330537+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000347511,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,0],"acting":[5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.0","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.012464+0000","last_change":"2026-04-17T12:19:42.012548+0000","last_active":"2026-04-17T12:19:42.012464+0000","last_peered":"2026-04-17T12:19:42.012464+0000","last_clean":"2026-04-17T12:19:42.012464+0000","last_became_active":"2026-04-17T12:19:40.024217+0000","last_became_peered":"2026-04-17T12:19:40.024217+0000","last_unstale":"2026-04-17T12:19:42.012464+0000","last_undegraded":"2026-04-17T12:19:42.012464+0000","last_fullsized":"2026-04-17T12:19:42.012464+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:51:06.772342+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00018744000000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]}],"pool_stats":[{"poolid":2,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2},{"poolid":1,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":8}],"osd_stats":[{"osd":7,"up_from":13,"seq":55834574851,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27624,"kb_used_data":632,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344216,"statfs":{"total":96636764160,"available":96608477184,"internally_reserved":0,"allocated":647168,"data_stored":517285,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6822,"internal_metadata":27452762},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":13,"seq":55834574851,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":26992,"kb_used_data":152,"kb_used_omap":5,"kb_used_meta":26810,"kb_avail":94344848,"statfs":{"total":96636764160,"available":96609124352,"internally_reserved":0,"allocated":155648,"data_stored":46065,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":5520,"internal_metadata":27454064},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":13,"seq":55834574851,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27024,"kb_used_data":176,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344816,"statfs":{"total":96636764160,"available":96609091584,"internally_reserved":0,"allocated":180224,"data_stored":57986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":13,"seq":55834574851,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27024,"kb_used_data":176,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344816,"statfs":{"total":96636764160,"available":96609091584,"internally_reserved":0,"allocated":180224,"data_stored":57986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6827,"internal_metadata":27452757},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":13,"seq":55834574851,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27024,"kb_used_data":176,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":94344816,"statfs":{"total":96636764160,"available":96609091584,"internally_reserved":0,"allocated":180224,"data_stored":57986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7475,"internal_metadata":27452109},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":13,"seq":55834574852,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27000,"kb_used_data":152,"kb_used_omap":5,"kb_used_meta":26810,"kb_avail":94344840,"statfs":{"total":96636764160,"available":96609116160,"internally_reserved":0,"allocated":155648,"data_stored":46065,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":5521,"internal_metadata":27454063},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574852,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27584,"kb_used_data":616,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344256,"statfs":{"total":96636764160,"available":96608518144,"internally_reserved":0,"allocated":630784,"data_stored":511450,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":13,"seq":55834574852,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27028,"kb_used_data":180,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":94344812,"statfs":{"total":96636764160,"available":96609087488,"internally_reserved":0,"allocated":184320,"data_stored":58005,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7477,"internal_metadata":27452107},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-04-17T12:19:46.528 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json 2026-04-17T12:19:46.736 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:46.736 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-04-17T12:19:46.751 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":16,"stamp":"2026-04-17T12:19:46.550438+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459299,"num_objects":4,"num_object_clones":0,"num_object_copies":8,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":4,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":59,"num_write_kb":586,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":35,"ondisk_log_size":35,"up":18,"acting":18,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":11,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":8,"kb":754974720,"kb_used":217300,"kb_used_data":2260,"kb_used_omap":50,"kb_used_meta":214477,"kb_avail":754757420,"statfs":{"total":773094113280,"available":772871598080,"internally_reserved":0,"allocated":2314240,"data_stored":1352828,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":51982,"internal_metadata":219624690},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":11,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":1,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.541991"},"pg_stats":[{"pgid":"1.7","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.713442+0000","last_change":"2026-04-17T12:19:42.713568+0000","last_active":"2026-04-17T12:19:42.713442+0000","last_peered":"2026-04-17T12:19:42.713442+0000","last_clean":"2026-04-17T12:19:42.713442+0000","last_became_active":"2026-04-17T12:19:40.013449+0000","last_became_peered":"2026-04-17T12:19:40.013449+0000","last_unstale":"2026-04-17T12:19:42.713442+0000","last_undegraded":"2026-04-17T12:19:42.713442+0000","last_fullsized":"2026-04-17T12:19:42.713442+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T15:47:10.055969+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000232876,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,3],"acting":[6,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.6","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.713419+0000","last_change":"2026-04-17T12:19:42.713542+0000","last_active":"2026-04-17T12:19:42.713419+0000","last_peered":"2026-04-17T12:19:42.713419+0000","last_clean":"2026-04-17T12:19:42.713419+0000","last_became_active":"2026-04-17T12:19:40.015415+0000","last_became_peered":"2026-04-17T12:19:40.015415+0000","last_unstale":"2026-04-17T12:19:42.713419+0000","last_undegraded":"2026-04-17T12:19:42.713419+0000","last_fullsized":"2026-04-17T12:19:42.713419+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T19:23:21.349991+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000284181,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[6,0],"acting":[6,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":6,"acting_primary":6,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.5","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.009970+0000","last_change":"2026-04-17T12:19:42.010374+0000","last_active":"2026-04-17T12:19:42.009970+0000","last_peered":"2026-04-17T12:19:42.009970+0000","last_clean":"2026-04-17T12:19:42.009970+0000","last_became_active":"2026-04-17T12:19:40.329751+0000","last_became_peered":"2026-04-17T12:19:40.329751+0000","last_unstale":"2026-04-17T12:19:42.009970+0000","last_undegraded":"2026-04-17T12:19:42.009970+0000","last_fullsized":"2026-04-17T12:19:42.009970+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:54:57.330537+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00051761799999999996,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[4,2],"acting":[4,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":4,"acting_primary":4,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.4","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.181017+0000","last_change":"2026-04-17T12:19:42.181110+0000","last_active":"2026-04-17T12:19:42.181017+0000","last_peered":"2026-04-17T12:19:42.181017+0000","last_clean":"2026-04-17T12:19:42.181017+0000","last_became_active":"2026-04-17T12:19:40.013650+0000","last_became_peered":"2026-04-17T12:19:40.013650+0000","last_unstale":"2026-04-17T12:19:42.181017+0000","last_undegraded":"2026-04-17T12:19:42.181017+0000","last_fullsized":"2026-04-17T12:19:42.181017+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T19:23:21.349991+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000293219,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5],"acting":[3,5],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"2.0","version":"16'32","reported_seq":59,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.012321+0000","last_change":"2026-04-17T12:19:40.013859+0000","last_active":"2026-04-17T12:19:42.012321+0000","last_peered":"2026-04-17T12:19:42.012321+0000","last_clean":"2026-04-17T12:19:42.012321+0000","last_became_active":"2026-04-17T12:19:40.013721+0000","last_became_peered":"2026-04-17T12:19:40.013721+0000","last_unstale":"2026-04-17T12:19:42.012321+0000","last_undegraded":"2026-04-17T12:19:42.012321+0000","last_fullsized":"2026-04-17T12:19:42.012321+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T13:22:59.361909+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,1],"acting":[7,1],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]},{"pgid":"1.3","version":"16'1","reported_seq":14,"reported_epoch":17,"state":"active+clean","last_fresh":"2026-04-17T12:19:41.009013+0000","last_change":"2026-04-17T12:19:40.014326+0000","last_active":"2026-04-17T12:19:41.009013+0000","last_peered":"2026-04-17T12:19:41.009013+0000","last_clean":"2026-04-17T12:19:41.009013+0000","last_became_active":"2026-04-17T12:19:40.013706+0000","last_became_peered":"2026-04-17T12:19:40.013706+0000","last_unstale":"2026-04-17T12:19:41.009013+0000","last_undegraded":"2026-04-17T12:19:41.009013+0000","last_fullsized":"2026-04-17T12:19:41.009013+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:54:57.330537+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[1,6],"acting":[1,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":1,"acting_primary":1,"purged_snaps":[]},{"pgid":"1.2","version":"18'2","reported_seq":22,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.013052+0000","last_change":"2026-04-17T12:19:42.013052+0000","last_active":"2026-04-17T12:19:42.013052+0000","last_peered":"2026-04-17T12:19:42.013052+0000","last_clean":"2026-04-17T12:19:42.013052+0000","last_became_active":"2026-04-17T12:19:40.015414+0000","last_became_peered":"2026-04-17T12:19:40.015414+0000","last_unstale":"2026-04-17T12:19:42.013052+0000","last_undegraded":"2026-04-17T12:19:42.013052+0000","last_fullsized":"2026-04-17T12:19:42.013052+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T13:03:17.584995+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00055161299999999995,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":2,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.1","version":"0'0","reported_seq":18,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.010176+0000","last_change":"2026-04-17T12:19:42.010320+0000","last_active":"2026-04-17T12:19:42.010176+0000","last_peered":"2026-04-17T12:19:42.010176+0000","last_clean":"2026-04-17T12:19:42.010176+0000","last_became_active":"2026-04-17T12:19:40.015608+0000","last_became_peered":"2026-04-17T12:19:40.015608+0000","last_unstale":"2026-04-17T12:19:42.010176+0000","last_undegraded":"2026-04-17T12:19:42.010176+0000","last_fullsized":"2026-04-17T12:19:42.010176+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:54:57.330537+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000347511,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[5,0],"acting":[5,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":5,"acting_primary":5,"purged_snaps":[{"start":"2","length":"1"}]},{"pgid":"1.0","version":"0'0","reported_seq":20,"reported_epoch":18,"state":"active+clean","last_fresh":"2026-04-17T12:19:42.012464+0000","last_change":"2026-04-17T12:19:42.012548+0000","last_active":"2026-04-17T12:19:42.012464+0000","last_peered":"2026-04-17T12:19:42.012464+0000","last_clean":"2026-04-17T12:19:42.012464+0000","last_became_active":"2026-04-17T12:19:40.024217+0000","last_became_peered":"2026-04-17T12:19:40.024217+0000","last_unstale":"2026-04-17T12:19:42.012464+0000","last_undegraded":"2026-04-17T12:19:42.012464+0000","last_fullsized":"2026-04-17T12:19:42.012464+0000","mapping_epoch":15,"log_start":"0'0","ondisk_log_start":"0'0","created":15,"last_epoch_clean":16,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-04-17T12:19:38.942247+0000","last_clean_scrub_stamp":"2026-04-17T12:19:38.942247+0000","objects_scrubbed":0,"log_size":0,"log_dups_size":0,"ondisk_log_size":0,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-04-18T12:51:06.772342+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00018744000000000001,"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,0],"acting":[7,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[{"start":"2","length":"1"}]}],"pool_stats":[{"poolid":2,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":925696,"data_stored":918560,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":2,"acting":2,"num_store_stats":2},{"poolid":1,"num_pg":8,"stat_sum":{"num_bytes":19,"num_objects":2,"num_object_clones":0,"num_object_copies":4,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":38,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":3,"ondisk_log_size":3,"up":16,"acting":16,"num_store_stats":8}],"osd_stats":[{"osd":7,"up_from":13,"seq":55834574851,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27624,"kb_used_data":632,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344216,"statfs":{"total":96636764160,"available":96608477184,"internally_reserved":0,"allocated":647168,"data_stored":517285,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6822,"internal_metadata":27452762},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":13,"seq":55834574851,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":26992,"kb_used_data":152,"kb_used_omap":5,"kb_used_meta":26810,"kb_avail":94344848,"statfs":{"total":96636764160,"available":96609124352,"internally_reserved":0,"allocated":155648,"data_stored":46065,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":5520,"internal_metadata":27454064},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":13,"seq":55834574851,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27024,"kb_used_data":176,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344816,"statfs":{"total":96636764160,"available":96609091584,"internally_reserved":0,"allocated":180224,"data_stored":57986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":13,"seq":55834574851,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27024,"kb_used_data":176,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344816,"statfs":{"total":96636764160,"available":96609091584,"internally_reserved":0,"allocated":180224,"data_stored":57986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6827,"internal_metadata":27452757},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":13,"seq":55834574851,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27024,"kb_used_data":176,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":94344816,"statfs":{"total":96636764160,"available":96609091584,"internally_reserved":0,"allocated":180224,"data_stored":57986,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7475,"internal_metadata":27452109},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":13,"seq":55834574852,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27000,"kb_used_data":152,"kb_used_omap":5,"kb_used_meta":26810,"kb_avail":94344840,"statfs":{"total":96636764160,"available":96609116160,"internally_reserved":0,"allocated":155648,"data_stored":46065,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":5521,"internal_metadata":27454063},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":1,"apply_latency_ms":1,"commit_latency_ns":1000000,"apply_latency_ns":1000000},"alerts":[]},{"osd":1,"up_from":13,"seq":55834574852,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27584,"kb_used_data":616,"kb_used_omap":6,"kb_used_meta":26809,"kb_avail":94344256,"statfs":{"total":96636764160,"available":96608518144,"internally_reserved":0,"allocated":630784,"data_stored":511450,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":6170,"internal_metadata":27453414},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":13,"seq":55834574852,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":94371840,"kb_used":27028,"kb_used_data":180,"kb_used_omap":7,"kb_used_meta":26808,"kb_avail":94344812,"statfs":{"total":96636764160,"available":96609087488,"internally_reserved":0,"allocated":184320,"data_stored":58005,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":7477,"internal_metadata":27452107},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":4,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-04-17T12:19:46.751 INFO:tasks.ceph.ceph_manager.ceph:clean! 2026-04-17T12:19:46.751 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-04-17T12:19:46.751 INFO:tasks.ceph.ceph_manager.ceph:wait_until_healthy 2026-04-17T12:19:46.751 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph health --format=json 2026-04-17T12:19:46.997 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:19:46.997 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-04-17T12:19:47.010 INFO:tasks.ceph.ceph_manager.ceph:wait_until_healthy done 2026-04-17T12:19:47.010 INFO:teuthology.run_tasks:Running task openssl_keys... 2026-04-17T12:19:47.014 INFO:teuthology.run_tasks:Running task rgw... 2026-04-17T12:19:47.019 DEBUG:tasks.rgw:config is {'client.0': None, 'client.1': None, 'client.2': None} 2026-04-17T12:19:47.019 DEBUG:tasks.rgw:client list is dict_keys(['client.0', 'client.1', 'client.2']) 2026-04-17T12:19:47.019 INFO:tasks.rgw:Creating data pools 2026-04-17T12:19:47.019 DEBUG:tasks.rgw:Obtaining remote for client client.0 2026-04-17T12:19:47.019 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster ceph 2026-04-17T12:19:48.048 INFO:teuthology.orchestra.run.vm02.stderr:pool 'default.rgw.buckets.data' created 2026-04-17T12:19:48.070 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster ceph 2026-04-17T12:19:49.055 INFO:teuthology.orchestra.run.vm02.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-04-17T12:19:49.090 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster ceph 2026-04-17T12:19:50.055 INFO:teuthology.orchestra.run.vm02.stderr:pool 'default.rgw.buckets.index' created 2026-04-17T12:19:50.081 DEBUG:teuthology.orchestra.run.vm02:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster ceph 2026-04-17T12:19:51.066 INFO:teuthology.orchestra.run.vm02.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-04-17T12:19:51.096 DEBUG:tasks.rgw:Obtaining remote for client client.1 2026-04-17T12:19:51.096 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster ceph 2026-04-17T12:19:51.302 INFO:teuthology.orchestra.run.vm06.stderr:pool 'default.rgw.buckets.data' already exists 2026-04-17T12:19:51.319 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster ceph 2026-04-17T12:19:52.064 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-04-17T12:19:52.076 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster ceph 2026-04-17T12:19:52.274 INFO:teuthology.orchestra.run.vm06.stderr:pool 'default.rgw.buckets.index' already exists 2026-04-17T12:19:52.287 DEBUG:teuthology.orchestra.run.vm06:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster ceph 2026-04-17T12:19:52.752 INFO:teuthology.orchestra.run.vm06.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-04-17T12:19:52.767 DEBUG:tasks.rgw:Obtaining remote for client client.2 2026-04-17T12:19:52.768 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create default.rgw.buckets.data 64 64 --cluster ceph 2026-04-17T12:19:52.981 INFO:teuthology.orchestra.run.vm08.stderr:pool 'default.rgw.buckets.data' already exists 2026-04-17T12:19:52.993 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable default.rgw.buckets.data rgw --cluster ceph 2026-04-17T12:19:54.099 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.data' 2026-04-17T12:19:54.113 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool create default.rgw.buckets.index 64 64 --cluster ceph 2026-04-17T12:19:54.317 INFO:teuthology.orchestra.run.vm08.stderr:pool 'default.rgw.buckets.index' already exists 2026-04-17T12:19:54.330 DEBUG:teuthology.orchestra.run.vm08:> sudo ceph osd pool application enable default.rgw.buckets.index rgw --cluster ceph 2026-04-17T12:19:55.109 INFO:teuthology.orchestra.run.vm08.stderr:enabled application 'rgw' on pool 'default.rgw.buckets.index' 2026-04-17T12:19:55.123 DEBUG:tasks.rgw:Pools created 2026-04-17T12:19:55.123 INFO:tasks.util.rgw:rgwadmin: client.0 : ['user', 'list'] 2026-04-17T12:19:55.124 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'user', 'list'] 2026-04-17T12:19:55.124 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph user list 2026-04-17T12:19:55.162 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:19:55.162 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:19:57.163 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.162+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.163+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.163+0000 7fe711414900 20 realm 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.163+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.163+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.163+0000 7fe711414900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.163+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.164+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:57.164 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.164+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.165 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.164+0000 7fe711414900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:19:57.165 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.164+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.166 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.165+0000 7fe711414900 20 rados_obj.operate() r=0 bl.length=1060 2026-04-17T12:19:57.166 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.165+0000 7fe711414900 20 searching for the correct realm 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 RGWRados::pool_iterate: got zonegroup_info.b9d64b71-9141-4977-a6f1-94a03af94bb6 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 RGWRados::pool_iterate: got zone_info.f9f3cc56-b810-479f-9b9d-92717e61b00d 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 RGWRados::pool_iterate: got default.zone. 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.174+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 rados_obj.operate() r=0 bl.length=436 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 zone default found 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 4 Realm: () 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 4 ZoneGroup: default (b9d64b71-9141-4977-a6f1-94a03af94bb6) 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 4 Zone: default (f9f3cc56-b810-479f-9b9d-92717e61b00d) 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 zonegroup default 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.175+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.176+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:57.176 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:57.176+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:59.148 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:59.147+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:59.148 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:59.147+0000 7fe711414900 20 rados->read ofs=0 len=0 2026-04-17T12:19:59.148 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:59.148+0000 7fe711414900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:19:59.148 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:59.148+0000 7fe711414900 20 started sync module instance, tier type = 2026-04-17T12:19:59.148 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:19:59.148+0000 7fe711414900 20 started zone id=f9f3cc56-b810-479f-9b9d-92717e61b00d (name=default) with tier type = 2026-04-17T12:20:01.196 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.195+0000 7fe711414900 20 add_watcher() i=0 2026-04-17T12:20:01.201 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.200+0000 7fe711414900 20 add_watcher() i=2 2026-04-17T12:20:01.201 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.200+0000 7fe711414900 20 add_watcher() i=4 2026-04-17T12:20:01.206 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.205+0000 7fe711414900 20 add_watcher() i=1 2026-04-17T12:20:01.208 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.207+0000 7fe711414900 20 add_watcher() i=3 2026-04-17T12:20:01.209 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.208+0000 7fe711414900 20 add_watcher() i=7 2026-04-17T12:20:01.210 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.209+0000 7fe711414900 20 add_watcher() i=6 2026-04-17T12:20:01.210 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.209+0000 7fe711414900 20 add_watcher() i=5 2026-04-17T12:20:01.210 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.209+0000 7fe711414900 2 all 8 watchers are set, enabling cache 2026-04-17T12:20:01.212 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.211+0000 7fe705ffb640 5 boost::asio::awaitable, obj_version> > logback_generations::read(const DoutPrefixProvider*):446: oid=data_loggenerations_metadata not found 2026-04-17T12:20:01.212 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.211+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.0 2026-04-17T12:20:01.212 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.211+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.213 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.212+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.0 does not exist 2026-04-17T12:20:01.213 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.212+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.1 2026-04-17T12:20:01.213 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.212+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.213 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.213+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.1 does not exist 2026-04-17T12:20:01.214 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.213+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.2 2026-04-17T12:20:01.214 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.213+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.214 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.213+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.2 does not exist 2026-04-17T12:20:01.214 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.213+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.3 2026-04-17T12:20:01.214 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.213+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.215 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.214+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.3 does not exist 2026-04-17T12:20:01.215 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.214+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.4 2026-04-17T12:20:01.215 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.214+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.215 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.214+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.4 does not exist 2026-04-17T12:20:01.215 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.214+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.5 2026-04-17T12:20:01.215 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.214+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.5 does not exist 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.6 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.6 does not exist 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.7 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.7 does not exist 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.8 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.215+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.8 does not exist 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.9 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.216 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.9 does not exist 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.10 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.10 does not exist 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.11 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.216+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.11 does not exist 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.12 2026-04-17T12:20:01.217 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.12 does not exist 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.13 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.13 does not exist 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.14 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.14 does not exist 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.15 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.15 does not exist 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.16 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.217+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.16 does not exist 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.17 2026-04-17T12:20:01.218 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.17 does not exist 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.18 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.18 does not exist 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.19 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.19 does not exist 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.20 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.20 does not exist 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.21 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.218+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.219 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.21 does not exist 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.22 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.22 does not exist 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.23 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.23 does not exist 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.24 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.24 does not exist 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.25 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.25 does not exist 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.26 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.219+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.26 does not exist 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.27 2026-04-17T12:20:01.220 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.27 does not exist 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.28 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.28 does not exist 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.29 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.29 does not exist 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.30 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.30 does not exist 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.31 2026-04-17T12:20:01.221 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.220+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.31 does not exist 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.32 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.32 does not exist 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.33 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.33 does not exist 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.34 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.34 does not exist 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.35 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.221+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.35 does not exist 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.36 2026-04-17T12:20:01.222 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.36 does not exist 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.37 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.37 does not exist 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.38 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.38 does not exist 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.39 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.222+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.39 does not exist 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.40 2026-04-17T12:20:01.223 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.40 does not exist 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.41 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.41 does not exist 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.42 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.42 does not exist 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.43 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.223+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.43 does not exist 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.44 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.224 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.44 does not exist 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.45 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.45 does not exist 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.46 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.46 does not exist 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.47 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.47 does not exist 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.48 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.224+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.48 does not exist 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.49 2026-04-17T12:20:01.225 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.49 does not exist 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.50 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.50 does not exist 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.51 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.51 does not exist 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.52 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.52 does not exist 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.53 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.225+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.53 does not exist 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.54 2026-04-17T12:20:01.226 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.54 does not exist 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.55 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.55 does not exist 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.56 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.56 does not exist 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.57 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.57 does not exist 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.58 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.226+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.58 does not exist 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.59 2026-04-17T12:20:01.227 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.59 does not exist 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.60 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.60 does not exist 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.61 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.61 does not exist 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.62 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.62 does not exist 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.63 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.227+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.63 does not exist 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.64 2026-04-17T12:20:01.228 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.64 does not exist 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.65 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.65 does not exist 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.66 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.66 does not exist 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.67 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.67 does not exist 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.68 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.228+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.68 does not exist 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.69 2026-04-17T12:20:01.229 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.69 does not exist 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.70 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.70 does not exist 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.71 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.71 does not exist 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.72 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.72 does not exist 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.73 2026-04-17T12:20:01.230 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.229+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.73 does not exist 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.74 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.74 does not exist 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.75 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.75 does not exist 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.76 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.76 does not exist 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.77 2026-04-17T12:20:01.231 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.230+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.77 does not exist 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.78 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.78 does not exist 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.79 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.79 does not exist 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.80 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.80 does not exist 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.81 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.231+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.232 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.81 does not exist 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.82 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.82 does not exist 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.83 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.83 does not exist 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.84 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.84 does not exist 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.85 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.232+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.85 does not exist 2026-04-17T12:20:01.233 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.86 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.86 does not exist 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.87 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.87 does not exist 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.88 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.88 does not exist 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.89 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.233+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.89 does not exist 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.90 2026-04-17T12:20:01.234 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.90 does not exist 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.91 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.91 does not exist 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.92 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.92 does not exist 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.93 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.234+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.93 does not exist 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.94 2026-04-17T12:20:01.235 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.94 does not exist 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.95 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.95 does not exist 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.96 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.96 does not exist 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.97 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.97 does not exist 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.98 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.235+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.98 does not exist 2026-04-17T12:20:01.236 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.99 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.99 does not exist 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.100 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.100 does not exist 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.101 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.101 does not exist 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.102 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.236+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.102 does not exist 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.103 2026-04-17T12:20:01.237 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.103 does not exist 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.104 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.104 does not exist 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.105 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.105 does not exist 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.106 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.106 does not exist 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.107 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.237+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.107 does not exist 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.108 2026-04-17T12:20:01.238 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.108 does not exist 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.109 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.109 does not exist 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.110 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.110 does not exist 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.111 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.111 does not exist 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.112 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.238+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.112 does not exist 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.113 2026-04-17T12:20:01.239 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.113 does not exist 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.114 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.114 does not exist 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.115 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.115 does not exist 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.116 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.116 does not exist 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.117 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.239+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.117 does not exist 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.118 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.118 does not exist 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.119 2026-04-17T12:20:01.240 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.119 does not exist 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.120 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe705ffb640 20 do_open: entering 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.120 does not exist 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe7057fa640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.121 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.121 does not exist 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe704ff9640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.122 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe704ff9640 20 do_open: entering 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.122 does not exist 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe70dbde640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.123 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.240+0000 7fe70dbde640 20 do_open: entering 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.123 does not exist 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe7077fe640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.124 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe7077fe640 20 do_open: entering 2026-04-17T12:20:01.241 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.124 does not exist 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe706ffd640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.125 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe706ffd640 20 do_open: entering 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.125 does not exist 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe7067fc640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.126 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe7067fc640 20 do_open: entering 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.126 does not exist 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe70ee67640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):59 probing obj=data_log.127 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe70ee67640 20 do_open: entering 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe705ffb640 20 boost::asio::awaitable {anonymous}::probe_shard(const DoutPrefixProvider*, neorados::RADOS, const neorados::Object&, const neorados::IOContext&, bool&):78: obj=data_log.127 does not exist 2026-04-17T12:20:01.242 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.241+0000 7fe705ffb640 20 do_create: entering 2026-04-17T12:20:01.243 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.243+0000 7fe7057fa640 20 do_open: entering 2026-04-17T12:20:01.246 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.245+0000 7fe711414900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:01.246 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:01.245+0000 7fe711414900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.252 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.251+0000 7fe711414900 10 rgw_init_ioctx warning: failed to set recovery_priority on default.rgw.meta 2026-04-17T12:20:04.252 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.252+0000 7fe711414900 5 note: GC not initialized 2026-04-17T12:20:04.253 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.252+0000 7fe6b57f2640 20 reqs_thread_entry: start 2026-04-17T12:20:04.315 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.314+0000 7fe711414900 20 init_complete bucket index max shards: 11 2026-04-17T12:20:04.315 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.314+0000 7fe711414900 20 Filter name: none 2026-04-17T12:20:04.315 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.314+0000 7fe6aeffd640 20 reqs_thread_entry: start 2026-04-17T12:20:04.325 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.324+0000 7fe711414900 20 remove_watcher() i=7 2026-04-17T12:20:04.325 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.324+0000 7fe711414900 2 removed watcher, disabling cache 2026-04-17T12:20:04.325 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.324+0000 7fe711414900 20 remove_watcher() i=0 2026-04-17T12:20:04.325 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.325+0000 7fe711414900 20 remove_watcher() i=3 2026-04-17T12:20:04.326 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.325+0000 7fe711414900 20 remove_watcher() i=5 2026-04-17T12:20:04.326 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.325+0000 7fe711414900 20 remove_watcher() i=6 2026-04-17T12:20:04.326 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.325+0000 7fe711414900 20 remove_watcher() i=2 2026-04-17T12:20:04.326 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.325+0000 7fe711414900 20 remove_watcher() i=4 2026-04-17T12:20:04.326 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.325+0000 7fe711414900 20 remove_watcher() i=1 2026-04-17T12:20:04.333 INFO:teuthology.orchestra.run.vm02.stdout:[] 2026-04-17T12:20:04.333 DEBUG:tasks.util.rgw: json result: [] 2026-04-17T12:20:04.333 INFO:tasks.rgw:Configuring storage class = FROZEN 2026-04-17T12:20:04.333 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-17T12:20:04.333 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-17T12:20:04.333 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN 2026-04-17T12:20:04.413 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:04.413 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:04.428 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.426+0000 7f8076d2c900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:04.428 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.426+0000 7f8076d2c900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.428 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.427+0000 7f8020ff1640 20 reqs_thread_entry: start 2026-04-17T12:20:04.437 INFO:teuthology.orchestra.run.vm02.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","STANDARD"]}}] 2026-04-17T12:20:04.437 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'STANDARD']}}] 2026-04-17T12:20:04.437 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-17T12:20:04.437 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-17T12:20:04.437 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN --data-pool default.rgw.buckets.data.frozen 2026-04-17T12:20:04.520 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:04.520 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:04.535 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.534+0000 7f98af739900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:04.535 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.534+0000 7f98af739900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.535 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.534+0000 7f9853fff640 20 reqs_thread_entry: start 2026-04-17T12:20:04.548 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"f9f3cc56-b810-479f-9b9d-92717e61b00d","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-17T12:20:04.548 DEBUG:tasks.util.rgw: json result: {'id': 'f9f3cc56-b810-479f-9b9d-92717e61b00d', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-17T12:20:04.548 INFO:tasks.rgw:Configuring storage class = LUKEWARM 2026-04-17T12:20:04.548 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-17T12:20:04.548 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-17T12:20:04.548 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM 2026-04-17T12:20:04.589 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:04.589 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:04.604 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.603+0000 7fab7ed23900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:04.604 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.603+0000 7fab7ed23900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.604 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.603+0000 7fab23fff640 20 reqs_thread_entry: start 2026-04-17T12:20:04.614 INFO:teuthology.orchestra.run.vm02.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-17T12:20:04.614 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-17T12:20:04.614 INFO:tasks.util.rgw:rgwadmin: client.0 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-17T12:20:04.614 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-17T12:20:04.614 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM --data-pool default.rgw.buckets.data.lukewarm 2026-04-17T12:20:04.701 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:04.701 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:04.717 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.716+0000 7fd74421b900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:04.718 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.716+0000 7fd74421b900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.718 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:04.717+0000 7fd6ed7f2640 20 reqs_thread_entry: start 2026-04-17T12:20:04.730 INFO:teuthology.orchestra.run.vm02.stdout:{"id":"f9f3cc56-b810-479f-9b9d-92717e61b00d","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-17T12:20:04.730 DEBUG:tasks.util.rgw: json result: {'id': 'f9f3cc56-b810-479f-9b9d-92717e61b00d', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-17T12:20:04.730 INFO:tasks.util.rgw:rgwadmin: client.1 : ['user', 'list'] 2026-04-17T12:20:04.730 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'user', 'list'] 2026-04-17T12:20:04.730 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph user list 2026-04-17T12:20:04.770 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:04.770 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:04.790 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.789+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 realm 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.791+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.792 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.792+0000 7fdf6cc10900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:20:04.793 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.792+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.793 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.792+0000 7fdf6cc10900 20 rados_obj.operate() r=0 bl.length=1190 2026-04-17T12:20:04.793 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.792+0000 7fdf6cc10900 20 searching for the correct realm 2026-04-17T12:20:04.803 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 RGWRados::pool_iterate: got zonegroup_info.b9d64b71-9141-4977-a6f1-94a03af94bb6 2026-04-17T12:20:04.803 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 RGWRados::pool_iterate: got zone_info.f9f3cc56-b810-479f-9b9d-92717e61b00d 2026-04-17T12:20:04.803 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-17T12:20:04.803 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 RGWRados::pool_iterate: got default.zone. 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.802+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados_obj.operate() r=0 bl.length=470 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 zone default found 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 4 Realm: () 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 4 ZoneGroup: default (b9d64b71-9141-4977-a6f1-94a03af94bb6) 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 4 Zone: default (f9f3cc56-b810-479f-9b9d-92717e61b00d) 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 zonegroup default 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.804 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.805 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.805 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.803+0000 7fdf6cc10900 20 rados->read ofs=0 len=0 2026-04-17T12:20:04.805 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.804+0000 7fdf6cc10900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:04.805 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.804+0000 7fdf6cc10900 20 started sync module instance, tier type = 2026-04-17T12:20:04.805 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.804+0000 7fdf6cc10900 20 started zone id=f9f3cc56-b810-479f-9b9d-92717e61b00d (name=default) with tier type = 2026-04-17T12:20:04.807 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.807+0000 7fdf6cc10900 20 add_watcher() i=0 2026-04-17T12:20:04.808 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.808+0000 7fdf6cc10900 20 add_watcher() i=1 2026-04-17T12:20:04.808 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.808+0000 7fdf6cc10900 20 add_watcher() i=3 2026-04-17T12:20:04.808 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.808+0000 7fdf6cc10900 20 add_watcher() i=4 2026-04-17T12:20:04.809 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.808+0000 7fdf6cc10900 20 add_watcher() i=5 2026-04-17T12:20:04.809 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.808+0000 7fdf6cc10900 20 add_watcher() i=2 2026-04-17T12:20:04.809 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.808+0000 7fdf6cc10900 20 add_watcher() i=6 2026-04-17T12:20:04.810 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.809+0000 7fdf6cc10900 20 add_watcher() i=7 2026-04-17T12:20:04.810 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.809+0000 7fdf6cc10900 2 all 8 watchers are set, enabling cache 2026-04-17T12:20:04.813 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.812+0000 7fdf6cc10900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:04.813 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.812+0000 7fdf6cc10900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.813 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.812+0000 7fdf6cc10900 5 note: GC not initialized 2026-04-17T12:20:04.813 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.812+0000 7fdf157ea640 20 reqs_thread_entry: start 2026-04-17T12:20:04.859 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.858+0000 7fdf6cc10900 20 init_complete bucket index max shards: 11 2026-04-17T12:20:04.859 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.858+0000 7fdf6cc10900 20 Filter name: none 2026-04-17T12:20:04.859 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.858+0000 7fdf0effd640 20 reqs_thread_entry: start 2026-04-17T12:20:04.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.870+0000 7fdf6cc10900 20 remove_watcher() i=0 2026-04-17T12:20:04.871 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.871+0000 7fdf6cc10900 2 removed watcher, disabling cache 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.871+0000 7fdf6cc10900 20 remove_watcher() i=7 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.871+0000 7fdf6cc10900 20 remove_watcher() i=5 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.871+0000 7fdf6cc10900 20 remove_watcher() i=3 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.871+0000 7fdf6cc10900 20 remove_watcher() i=6 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.872+0000 7fdf6cc10900 20 remove_watcher() i=1 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.872+0000 7fdf6cc10900 20 remove_watcher() i=4 2026-04-17T12:20:04.872 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.872+0000 7fdf6cc10900 20 remove_watcher() i=2 2026-04-17T12:20:04.881 INFO:teuthology.orchestra.run.vm06.stdout:[] 2026-04-17T12:20:04.881 DEBUG:tasks.util.rgw: json result: [] 2026-04-17T12:20:04.881 INFO:tasks.rgw:Configuring storage class = FROZEN 2026-04-17T12:20:04.881 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-17T12:20:04.881 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-17T12:20:04.881 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN 2026-04-17T12:20:04.966 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:04.966 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:04.981 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.980+0000 7f8580f20900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:04.981 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.980+0000 7f8580f20900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:04.981 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:04.981+0000 7f8529ff3640 20 reqs_thread_entry: start 2026-04-17T12:20:04.992 INFO:teuthology.orchestra.run.vm06.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-17T12:20:04.992 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-17T12:20:04.992 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-17T12:20:04.992 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-17T12:20:04.992 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN --data-pool default.rgw.buckets.data.frozen 2026-04-17T12:20:05.035 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.035 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.049 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.048+0000 7f468b01b900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.049 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.048+0000 7f468b01b900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.049 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.049+0000 7f4633fef640 20 reqs_thread_entry: start 2026-04-17T12:20:05.061 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"f9f3cc56-b810-479f-9b9d-92717e61b00d","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-17T12:20:05.061 DEBUG:tasks.util.rgw: json result: {'id': 'f9f3cc56-b810-479f-9b9d-92717e61b00d', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-17T12:20:05.061 INFO:tasks.rgw:Configuring storage class = LUKEWARM 2026-04-17T12:20:05.061 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-17T12:20:05.061 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-17T12:20:05.061 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM 2026-04-17T12:20:05.102 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.102 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.119 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.118+0000 7febdf610900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.119 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.118+0000 7febdf610900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.119 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.118+0000 7feb83fff640 20 reqs_thread_entry: start 2026-04-17T12:20:05.128 INFO:teuthology.orchestra.run.vm06.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-17T12:20:05.128 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-17T12:20:05.129 INFO:tasks.util.rgw:rgwadmin: client.1 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-17T12:20:05.129 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.1', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-17T12:20:05.129 DEBUG:teuthology.orchestra.run.vm06:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.1 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM --data-pool default.rgw.buckets.data.lukewarm 2026-04-17T12:20:05.209 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.210 INFO:teuthology.orchestra.run.vm06.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.224 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.222+0000 7f605c752900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.224 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.223+0000 7f605c752900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.224 INFO:teuthology.orchestra.run.vm06.stderr:2026-04-17T12:20:05.223+0000 7f60037fe640 20 reqs_thread_entry: start 2026-04-17T12:20:05.234 INFO:teuthology.orchestra.run.vm06.stdout:{"id":"f9f3cc56-b810-479f-9b9d-92717e61b00d","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-17T12:20:05.235 DEBUG:tasks.util.rgw: json result: {'id': 'f9f3cc56-b810-479f-9b9d-92717e61b00d', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-17T12:20:05.235 INFO:tasks.util.rgw:rgwadmin: client.2 : ['user', 'list'] 2026-04-17T12:20:05.235 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'user', 'list'] 2026-04-17T12:20:05.235 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph user list 2026-04-17T12:20:05.272 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.272 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.289 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.288+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 20 realm 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.289+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.291 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.290+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.292 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.290+0000 7f774a34d900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:20:05.292 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.290+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.292 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.291+0000 7f774a34d900 20 rados_obj.operate() r=0 bl.length=1190 2026-04-17T12:20:05.292 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.291+0000 7f774a34d900 20 searching for the correct realm 2026-04-17T12:20:05.301 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 RGWRados::pool_iterate: got zonegroup_info.b9d64b71-9141-4977-a6f1-94a03af94bb6 2026-04-17T12:20:05.301 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 RGWRados::pool_iterate: got zone_info.f9f3cc56-b810-479f-9b9d-92717e61b00d 2026-04-17T12:20:05.301 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-17T12:20:05.301 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 RGWRados::pool_iterate: got default.zone. 2026-04-17T12:20:05.301 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-17T12:20:05.301 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.299+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados_obj.operate() r=0 bl.length=470 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 zone default found 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 4 Realm: () 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 4 ZoneGroup: default (b9d64b71-9141-4977-a6f1-94a03af94bb6) 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 4 Zone: default (f9f3cc56-b810-479f-9b9d-92717e61b00d) 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 zonegroup default 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.302 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.300+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.303 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.301+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.303 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.301+0000 7f774a34d900 20 rados->read ofs=0 len=0 2026-04-17T12:20:05.303 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.301+0000 7f774a34d900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:05.303 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.301+0000 7f774a34d900 20 started sync module instance, tier type = 2026-04-17T12:20:05.303 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.301+0000 7f774a34d900 20 started zone id=f9f3cc56-b810-479f-9b9d-92717e61b00d (name=default) with tier type = 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.304+0000 7f774a34d900 20 add_watcher() i=2 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.304+0000 7f774a34d900 20 add_watcher() i=0 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.304+0000 7f774a34d900 20 add_watcher() i=5 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.304+0000 7f774a34d900 20 add_watcher() i=6 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.305+0000 7f774a34d900 20 add_watcher() i=1 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.305+0000 7f774a34d900 20 add_watcher() i=4 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.305+0000 7f774a34d900 20 add_watcher() i=3 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.305+0000 7f774a34d900 20 add_watcher() i=7 2026-04-17T12:20:05.306 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.305+0000 7f774a34d900 2 all 8 watchers are set, enabling cache 2026-04-17T12:20:05.309 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.307+0000 7f774a34d900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.309 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.307+0000 7f774a34d900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.309 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.307+0000 7f774a34d900 5 note: GC not initialized 2026-04-17T12:20:05.310 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.308+0000 7f76eeffd640 20 reqs_thread_entry: start 2026-04-17T12:20:05.354 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.352+0000 7f774a34d900 20 init_complete bucket index max shards: 11 2026-04-17T12:20:05.354 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.352+0000 7f774a34d900 20 Filter name: none 2026-04-17T12:20:05.354 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.352+0000 7f76ecff9640 20 reqs_thread_entry: start 2026-04-17T12:20:05.363 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.362+0000 7f774a34d900 20 remove_watcher() i=5 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.362+0000 7f774a34d900 2 removed watcher, disabling cache 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.362+0000 7f774a34d900 20 remove_watcher() i=3 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.362+0000 7f774a34d900 20 remove_watcher() i=0 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.363+0000 7f774a34d900 20 remove_watcher() i=4 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.363+0000 7f774a34d900 20 remove_watcher() i=2 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.363+0000 7f774a34d900 20 remove_watcher() i=1 2026-04-17T12:20:05.364 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.363+0000 7f774a34d900 20 remove_watcher() i=6 2026-04-17T12:20:05.365 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.363+0000 7f774a34d900 20 remove_watcher() i=7 2026-04-17T12:20:05.371 INFO:teuthology.orchestra.run.vm08.stdout:[] 2026-04-17T12:20:05.371 DEBUG:tasks.util.rgw: json result: [] 2026-04-17T12:20:05.371 INFO:tasks.rgw:Configuring storage class = FROZEN 2026-04-17T12:20:05.371 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-17T12:20:05.371 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN'] 2026-04-17T12:20:05.371 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN 2026-04-17T12:20:05.451 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.452 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.467 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.466+0000 7f4ec1b38900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.467 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.466+0000 7f4ec1b38900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.468 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.466+0000 7f4e6b7ee640 20 reqs_thread_entry: start 2026-04-17T12:20:05.479 INFO:teuthology.orchestra.run.vm08.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-17T12:20:05.479 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-17T12:20:05.479 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-17T12:20:05.479 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'FROZEN', '--data-pool', 'default.rgw.buckets.data.frozen'] 2026-04-17T12:20:05.479 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class FROZEN --data-pool default.rgw.buckets.data.frozen 2026-04-17T12:20:05.563 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.563 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.579 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.577+0000 7fe087ba3900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.579 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.577+0000 7fe087ba3900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.579 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.577+0000 7fe030ff1640 20 reqs_thread_entry: start 2026-04-17T12:20:05.591 INFO:teuthology.orchestra.run.vm08.stdout:{"id":"f9f3cc56-b810-479f-9b9d-92717e61b00d","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-17T12:20:05.591 DEBUG:tasks.util.rgw: json result: {'id': 'f9f3cc56-b810-479f-9b9d-92717e61b00d', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-17T12:20:05.591 INFO:tasks.rgw:Configuring storage class = LUKEWARM 2026-04-17T12:20:05.591 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-17T12:20:05.591 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zonegroup', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM'] 2026-04-17T12:20:05.591 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zonegroup placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM 2026-04-17T12:20:05.634 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.634 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.651 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.650+0000 7f7ef6f32900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.651 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.650+0000 7f7ef6f32900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.652 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.650+0000 7f7ea0ff1640 20 reqs_thread_entry: start 2026-04-17T12:20:05.663 INFO:teuthology.orchestra.run.vm08.stdout:[{"key":"default-placement","val":{"name":"default-placement","tags":[],"storage_classes":["FROZEN","LUKEWARM","STANDARD"]}}] 2026-04-17T12:20:05.663 DEBUG:tasks.util.rgw: json result: [{'key': 'default-placement', 'val': {'name': 'default-placement', 'tags': [], 'storage_classes': ['FROZEN', 'LUKEWARM', 'STANDARD']}}] 2026-04-17T12:20:05.663 INFO:tasks.util.rgw:rgwadmin: client.2 : ['zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-17T12:20:05.663 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.2', '--cluster', 'ceph', 'zone', 'placement', 'add', '--rgw-zone', 'default', '--placement-id', 'default-placement', '--storage-class', 'LUKEWARM', '--data-pool', 'default.rgw.buckets.data.lukewarm'] 2026-04-17T12:20:05.663 DEBUG:teuthology.orchestra.run.vm08:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.2 --cluster ceph zone placement add --rgw-zone default --placement-id default-placement --storage-class LUKEWARM --data-pool default.rgw.buckets.data.lukewarm 2026-04-17T12:20:05.707 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:05.707 INFO:teuthology.orchestra.run.vm08.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:05.723 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.721+0000 7fca4b412900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:05.723 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.721+0000 7fca4b412900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:05.723 INFO:teuthology.orchestra.run.vm08.stderr:2026-04-17T12:20:05.721+0000 7fc9effff640 20 reqs_thread_entry: start 2026-04-17T12:20:05.735 INFO:teuthology.orchestra.run.vm08.stdout:{"id":"f9f3cc56-b810-479f-9b9d-92717e61b00d","name":"default","domain_root":"default.rgw.meta:root","control_pool":"default.rgw.control","dedup_pool":"default.rgw.dedup","gc_pool":"default.rgw.log:gc","lc_pool":"default.rgw.log:lc","log_pool":"default.rgw.log","intent_log_pool":"default.rgw.log:intent","usage_log_pool":"default.rgw.log:usage","roles_pool":"default.rgw.meta:roles","reshard_pool":"default.rgw.log:reshard","user_keys_pool":"default.rgw.meta:users.keys","user_email_pool":"default.rgw.meta:users.email","user_swift_pool":"default.rgw.meta:users.swift","user_uid_pool":"default.rgw.meta:users.uid","otp_pool":"default.rgw.otp","notif_pool":"default.rgw.log:notif","topics_pool":"default.rgw.meta:topics","account_pool":"default.rgw.meta:accounts","group_pool":"default.rgw.meta:groups","system_key":{"access_key":"","secret_key":""},"placement_pools":[{"key":"default-placement","val":{"index_pool":"default.rgw.buckets.index","storage_classes":{"FROZEN":{"data_pool":"default.rgw.buckets.data.frozen"},"LUKEWARM":{"data_pool":"default.rgw.buckets.data.lukewarm"},"STANDARD":{"data_pool":"default.rgw.buckets.data"}},"data_extra_pool":"default.rgw.buckets.non-ec","index_type":0,"inline_data":true}}],"realm_id":"","restore_pool":"default.rgw.log:restore"} 2026-04-17T12:20:05.735 DEBUG:tasks.util.rgw: json result: {'id': 'f9f3cc56-b810-479f-9b9d-92717e61b00d', 'name': 'default', 'domain_root': 'default.rgw.meta:root', 'control_pool': 'default.rgw.control', 'dedup_pool': 'default.rgw.dedup', 'gc_pool': 'default.rgw.log:gc', 'lc_pool': 'default.rgw.log:lc', 'log_pool': 'default.rgw.log', 'intent_log_pool': 'default.rgw.log:intent', 'usage_log_pool': 'default.rgw.log:usage', 'roles_pool': 'default.rgw.meta:roles', 'reshard_pool': 'default.rgw.log:reshard', 'user_keys_pool': 'default.rgw.meta:users.keys', 'user_email_pool': 'default.rgw.meta:users.email', 'user_swift_pool': 'default.rgw.meta:users.swift', 'user_uid_pool': 'default.rgw.meta:users.uid', 'otp_pool': 'default.rgw.otp', 'notif_pool': 'default.rgw.log:notif', 'topics_pool': 'default.rgw.meta:topics', 'account_pool': 'default.rgw.meta:accounts', 'group_pool': 'default.rgw.meta:groups', 'system_key': {'access_key': '', 'secret_key': ''}, 'placement_pools': [{'key': 'default-placement', 'val': {'index_pool': 'default.rgw.buckets.index', 'storage_classes': {'FROZEN': {'data_pool': 'default.rgw.buckets.data.frozen'}, 'LUKEWARM': {'data_pool': 'default.rgw.buckets.data.lukewarm'}, 'STANDARD': {'data_pool': 'default.rgw.buckets.data'}}, 'data_extra_pool': 'default.rgw.buckets.non-ec', 'index_type': 0, 'inline_data': True}}], 'realm_id': '', 'restore_pool': 'default.rgw.log:restore'} 2026-04-17T12:20:05.736 INFO:tasks.rgw:Starting rgw... 2026-04-17T12:20:05.736 INFO:tasks.rgw:rgw client.0 config is {} 2026-04-17T12:20:05.736 INFO:tasks.rgw:Using beast as radosgw frontend 2026-04-17T12:20:05.736 DEBUG:teuthology.orchestra.run.vm02:> sudo echo -n http://vm02.local:80 | sudo tee /home/ubuntu/cephtest/url_file 2026-04-17T12:20:05.765 INFO:teuthology.orchestra.run.vm02.stdout:http://vm02.local:80 2026-04-17T12:20:05.765 DEBUG:teuthology.orchestra.run.vm02:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-04-17T12:20:05.833 INFO:tasks.rgw.client.0:Restarting daemon 2026-04-17T12:20:05.833 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.0 --cluster ceph -k /etc/ceph/ceph.client.0.keyring --log-file /var/log/ceph/rgw.ceph.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.0.stdout 2>&1 2026-04-17T12:20:05.874 INFO:tasks.rgw.client.0:Started 2026-04-17T12:20:05.874 INFO:tasks.rgw:rgw client.1 config is {} 2026-04-17T12:20:05.874 INFO:tasks.rgw:Using beast as radosgw frontend 2026-04-17T12:20:05.874 DEBUG:teuthology.orchestra.run.vm06:> sudo echo -n http://vm06.local:80 | sudo tee /home/ubuntu/cephtest/url_file 2026-04-17T12:20:05.903 INFO:teuthology.orchestra.run.vm06.stdout:http://vm06.local:80 2026-04-17T12:20:05.903 DEBUG:teuthology.orchestra.run.vm06:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-04-17T12:20:05.969 INFO:tasks.rgw.client.1:Restarting daemon 2026-04-17T12:20:05.969 DEBUG:teuthology.orchestra.run.vm06:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.1 --cluster ceph -k /etc/ceph/ceph.client.1.keyring --log-file /var/log/ceph/rgw.ceph.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.1.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.1.stdout 2>&1 2026-04-17T12:20:06.011 INFO:tasks.rgw.client.1:Started 2026-04-17T12:20:06.011 INFO:tasks.rgw:rgw client.2 config is {} 2026-04-17T12:20:06.011 INFO:tasks.rgw:Using beast as radosgw frontend 2026-04-17T12:20:06.011 DEBUG:teuthology.orchestra.run.vm08:> sudo echo -n http://vm08.local:80 | sudo tee /home/ubuntu/cephtest/url_file 2026-04-17T12:20:06.041 INFO:teuthology.orchestra.run.vm08.stdout:http://vm08.local:80 2026-04-17T12:20:06.041 DEBUG:teuthology.orchestra.run.vm08:> sudo chown ceph /home/ubuntu/cephtest/url_file 2026-04-17T12:20:06.107 INFO:tasks.rgw.client.2:Restarting daemon 2026-04-17T12:20:06.107 DEBUG:teuthology.orchestra.run.vm08:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.2 --cluster ceph -k /etc/ceph/ceph.client.2.keyring --log-file /var/log/ceph/rgw.ceph.client.2.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.2.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.2.stdout 2>&1 2026-04-17T12:20:06.150 INFO:tasks.rgw.client.2:Started 2026-04-17T12:20:06.150 INFO:tasks.rgw:Polling client.0 until it starts accepting connections on http://vm02.local:80/ 2026-04-17T12:20:06.150 DEBUG:teuthology.orchestra.run.vm02:> curl http://vm02.local:80/ 2026-04-17T12:20:06.182 INFO:teuthology.orchestra.run.vm02.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-17T12:20:06.182 INFO:teuthology.orchestra.run.vm02.stderr: Dload Upload Total Spent Left Speed 2026-04-17T12:20:06.186 INFO:teuthology.orchestra.run.vm02.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 46750 0 --:--:-- --:--:-- --:--:-- 46750 2026-04-17T12:20:06.187 INFO:teuthology.orchestra.run.vm02.stdout:anonymous 2026-04-17T12:20:06.187 INFO:tasks.rgw:Polling client.1 until it starts accepting connections on http://vm06.local:80/ 2026-04-17T12:20:06.187 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:80/ 2026-04-17T12:20:06.212 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-17T12:20:06.212 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-04-17T12:20:06.212 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 2026-04-17T12:20:06.212 INFO:teuthology.orchestra.run.vm06.stderr:curl: (7) Failed to connect to vm06.local port 80: Connection refused 2026-04-17T12:20:06.213 DEBUG:teuthology.orchestra.run:got remote process result: 7 2026-04-17T12:20:07.213 DEBUG:teuthology.orchestra.run.vm06:> curl http://vm06.local:80/ 2026-04-17T12:20:07.234 INFO:teuthology.orchestra.run.vm06.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-17T12:20:07.235 INFO:teuthology.orchestra.run.vm06.stderr: Dload Upload Total Spent Left Speed 2026-04-17T12:20:07.236 INFO:teuthology.orchestra.run.vm06.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-04-17T12:20:07.237 INFO:teuthology.orchestra.run.vm06.stdout:anonymous 2026-04-17T12:20:07.237 INFO:tasks.rgw:Polling client.2 until it starts accepting connections on http://vm08.local:80/ 2026-04-17T12:20:07.237 DEBUG:teuthology.orchestra.run.vm08:> curl http://vm08.local:80/ 2026-04-17T12:20:07.256 INFO:teuthology.orchestra.run.vm08.stderr: % Total % Received % Xferd Average Speed Time Time Time Current 2026-04-17T12:20:07.256 INFO:teuthology.orchestra.run.vm08.stderr: Dload Upload Total Spent Left Speed 2026-04-17T12:20:07.257 INFO:teuthology.orchestra.run.vm08.stderr: 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 187 0 187 0 0 182k 0 --:--:-- --:--:-- --:--:-- 182k 2026-04-17T12:20:07.257 INFO:teuthology.orchestra.run.vm08.stdout:anonymous 2026-04-17T12:20:07.258 INFO:teuthology.run_tasks:Running task tox... 2026-04-17T12:20:07.260 INFO:tasks.tox:Deploying tox from pip... 2026-04-17T12:20:07.260 DEBUG:teuthology.orchestra.run.vm02:> python3 -m venv /home/ubuntu/cephtest/tox-venv 2026-04-17T12:20:08.650 DEBUG:teuthology.orchestra.run.vm02:> source /home/ubuntu/cephtest/tox-venv/bin/activate && pip install tox 2026-04-17T12:20:08.957 INFO:teuthology.orchestra.run.vm02.stdout:Collecting tox 2026-04-17T12:20:08.986 INFO:teuthology.orchestra.run.vm02.stdout: Downloading tox-4.30.3-py3-none-any.whl (175 kB) 2026-04-17T12:20:09.075 INFO:teuthology.orchestra.run.vm02.stdout:Collecting chardet>=5.2 2026-04-17T12:20:09.084 INFO:teuthology.orchestra.run.vm02.stdout: Downloading chardet-5.2.0-py3-none-any.whl (199 kB) 2026-04-17T12:20:09.112 INFO:teuthology.orchestra.run.vm02.stdout:Collecting colorama>=0.4.6 2026-04-17T12:20:09.122 INFO:teuthology.orchestra.run.vm02.stdout: Downloading colorama-0.4.6-py2.py3-none-any.whl (25 kB) 2026-04-17T12:20:09.153 INFO:teuthology.orchestra.run.vm02.stdout:Collecting cachetools>=6.1 2026-04-17T12:20:09.162 INFO:teuthology.orchestra.run.vm02.stdout: Downloading cachetools-6.2.6-py3-none-any.whl (11 kB) 2026-04-17T12:20:09.201 INFO:teuthology.orchestra.run.vm02.stdout:Collecting filelock>=3.18 2026-04-17T12:20:09.211 INFO:teuthology.orchestra.run.vm02.stdout: Downloading filelock-3.19.1-py3-none-any.whl (15 kB) 2026-04-17T12:20:09.245 INFO:teuthology.orchestra.run.vm02.stdout:Collecting platformdirs>=4.3.8 2026-04-17T12:20:09.254 INFO:teuthology.orchestra.run.vm02.stdout: Downloading platformdirs-4.4.0-py3-none-any.whl (18 kB) 2026-04-17T12:20:09.285 INFO:teuthology.orchestra.run.vm02.stdout:Collecting packaging>=25 2026-04-17T12:20:09.293 INFO:teuthology.orchestra.run.vm02.stdout: Downloading packaging-26.1-py3-none-any.whl (95 kB) 2026-04-17T12:20:09.321 INFO:teuthology.orchestra.run.vm02.stdout:Collecting pyproject-api>=1.9.1 2026-04-17T12:20:09.331 INFO:teuthology.orchestra.run.vm02.stdout: Downloading pyproject_api-1.9.1-py3-none-any.whl (13 kB) 2026-04-17T12:20:09.365 INFO:teuthology.orchestra.run.vm02.stdout:Collecting typing-extensions>=4.14.1 2026-04-17T12:20:09.373 INFO:teuthology.orchestra.run.vm02.stdout: Downloading typing_extensions-4.15.0-py3-none-any.whl (44 kB) 2026-04-17T12:20:09.397 INFO:teuthology.orchestra.run.vm02.stdout:Collecting pluggy>=1.6 2026-04-17T12:20:09.406 INFO:teuthology.orchestra.run.vm02.stdout: Downloading pluggy-1.6.0-py3-none-any.whl (20 kB) 2026-04-17T12:20:09.460 INFO:teuthology.orchestra.run.vm02.stdout:Collecting tomli>=2.2.1 2026-04-17T12:20:09.469 INFO:teuthology.orchestra.run.vm02.stdout: Downloading tomli-2.4.1-py3-none-any.whl (14 kB) 2026-04-17T12:20:09.561 INFO:teuthology.orchestra.run.vm02.stdout:Collecting virtualenv>=20.31.2 2026-04-17T12:20:09.570 INFO:teuthology.orchestra.run.vm02.stdout: Downloading virtualenv-21.2.4-py3-none-any.whl (5.8 MB) 2026-04-17T12:20:09.688 INFO:teuthology.orchestra.run.vm02.stdout:Collecting distlib<1,>=0.3.7 2026-04-17T12:20:09.696 INFO:teuthology.orchestra.run.vm02.stdout: Downloading distlib-0.4.0-py2.py3-none-any.whl (469 kB) 2026-04-17T12:20:09.723 INFO:teuthology.orchestra.run.vm02.stdout:Collecting python-discovery>=1.2.2 2026-04-17T12:20:09.731 INFO:teuthology.orchestra.run.vm02.stdout: Downloading python_discovery-1.2.2-py3-none-any.whl (31 kB) 2026-04-17T12:20:09.792 INFO:teuthology.orchestra.run.vm02.stdout:Installing collected packages: platformdirs, filelock, typing-extensions, tomli, python-discovery, packaging, distlib, virtualenv, pyproject-api, pluggy, colorama, chardet, cachetools, tox 2026-04-17T12:20:10.163 INFO:teuthology.orchestra.run.vm02.stdout:Successfully installed cachetools-6.2.6 chardet-5.2.0 colorama-0.4.6 distlib-0.4.0 filelock-3.19.1 packaging-26.1 platformdirs-4.4.0 pluggy-1.6.0 pyproject-api-1.9.1 python-discovery-1.2.2 tomli-2.4.1 tox-4.30.3 typing-extensions-4.15.0 virtualenv-21.2.4 2026-04-17T12:20:10.240 INFO:teuthology.orchestra.run.vm02.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-04-17T12:20:10.240 INFO:teuthology.orchestra.run.vm02.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/tox-venv/bin/python3 -m pip install --upgrade pip' command. 2026-04-17T12:20:10.296 INFO:teuthology.run_tasks:Running task tox... 2026-04-17T12:20:10.298 INFO:tasks.tox:Deploying tox from pip... 2026-04-17T12:20:10.298 DEBUG:teuthology.orchestra.run.vm02:> python3 -m venv /home/ubuntu/cephtest/tox-venv 2026-04-17T12:20:11.039 DEBUG:teuthology.orchestra.run.vm02:> source /home/ubuntu/cephtest/tox-venv/bin/activate && pip install tox 2026-04-17T12:20:11.191 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: tox in ./cephtest/tox-venv/lib/python3.9/site-packages (4.30.3) 2026-04-17T12:20:11.196 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: chardet>=5.2 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (5.2.0) 2026-04-17T12:20:11.196 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: cachetools>=6.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (6.2.6) 2026-04-17T12:20:11.196 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: colorama>=0.4.6 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (0.4.6) 2026-04-17T12:20:11.196 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: platformdirs>=4.3.8 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (4.4.0) 2026-04-17T12:20:11.197 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: filelock>=3.18 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (3.19.1) 2026-04-17T12:20:11.197 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: pyproject-api>=1.9.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (1.9.1) 2026-04-17T12:20:11.197 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: tomli>=2.2.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (2.4.1) 2026-04-17T12:20:11.198 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: typing-extensions>=4.14.1 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (4.15.0) 2026-04-17T12:20:11.198 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: virtualenv>=20.31.2 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (21.2.4) 2026-04-17T12:20:11.198 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: packaging>=25 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (26.1) 2026-04-17T12:20:11.198 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: pluggy>=1.6 in ./cephtest/tox-venv/lib/python3.9/site-packages (from tox) (1.6.0) 2026-04-17T12:20:11.222 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: python-discovery>=1.2.2 in ./cephtest/tox-venv/lib/python3.9/site-packages (from virtualenv>=20.31.2->tox) (1.2.2) 2026-04-17T12:20:11.222 INFO:teuthology.orchestra.run.vm02.stdout:Requirement already satisfied: distlib<1,>=0.3.7 in ./cephtest/tox-venv/lib/python3.9/site-packages (from virtualenv>=20.31.2->tox) (0.4.0) 2026-04-17T12:20:11.238 INFO:teuthology.orchestra.run.vm02.stderr:WARNING: You are using pip version 21.3.1; however, version 26.0.1 is available. 2026-04-17T12:20:11.238 INFO:teuthology.orchestra.run.vm02.stderr:You should consider upgrading via the '/home/ubuntu/cephtest/tox-venv/bin/python3 -m pip install --upgrade pip' command. 2026-04-17T12:20:11.270 INFO:teuthology.run_tasks:Running task dedup-tests... 2026-04-17T12:20:11.274 DEBUG:tasks.dedup_tests:config is {'client.0': {'rgw_server': 'client.0'}} 2026-04-17T12:20:11.274 INFO:tasks.dedup_tests:Downloading dedup-tests... 2026-04-17T12:20:11.274 INFO:tasks.dedup_tests:Using branch tt-wip-sse-s3-on-v20.2.0 from http://git.local/ceph.git for dedup tests 2026-04-17T12:20:11.274 DEBUG:teuthology.orchestra.run.vm02:> git clone -b tt-wip-sse-s3-on-v20.2.0 http://git.local/ceph.git /home/ubuntu/cephtest/ceph 2026-04-17T12:20:11.292 INFO:teuthology.orchestra.run.vm02.stderr:Cloning into '/home/ubuntu/cephtest/ceph'... 2026-04-17T12:20:43.771 INFO:tasks.dedup_tests:Creating rgw user... 2026-04-17T12:20:43.771 DEBUG:tasks.dedup_tests:Creating user foo.client.0 on client.0 2026-04-17T12:20:43.771 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user create --uid foo.client.0 --display-name 'Mr. foo.client.0' --access-key WZNRJECYHIJXXDDQHDUU --secret LVqqWb6ixrnFJANAuw7aL+nM3sZlqbJRogGtsy3lZjDz6pzRhxNPXw== --cluster ceph 2026-04-17T12:20:43.854 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:20:43.854 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:20:43.874 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.873+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 realm 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 4 RGWPeriod::init failed to init realm id : (2) No such file or directory 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.875 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.874+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.876 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.875+0000 7f0a977a3900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:20:43.876 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.875+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.877 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.876+0000 7f0a977a3900 20 rados_obj.operate() r=0 bl.length=1190 2026-04-17T12:20:43.877 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.876+0000 7f0a977a3900 20 searching for the correct realm 2026-04-17T12:20:43.885 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 RGWRados::pool_iterate: got zonegroup_info.b9d64b71-9141-4977-a6f1-94a03af94bb6 2026-04-17T12:20:43.885 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 RGWRados::pool_iterate: got zone_info.f9f3cc56-b810-479f-9b9d-92717e61b00d 2026-04-17T12:20:43.885 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 RGWRados::pool_iterate: got default.zonegroup. 2026-04-17T12:20:43.885 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 RGWRados::pool_iterate: got default.zone. 2026-04-17T12:20:43.885 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 RGWRados::pool_iterate: got zone_names.default 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 RGWRados::pool_iterate: got zonegroups_names.default 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.884+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados_obj.operate() r=0 bl.length=46 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados_obj.operate() r=0 bl.length=470 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 zone default found 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 4 Realm: () 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 4 ZoneGroup: default (b9d64b71-9141-4977-a6f1-94a03af94bb6) 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 4 Zone: default (f9f3cc56-b810-479f-9b9d-92717e61b00d) 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 10 cannot find current period zonegroup using local zonegroup configuration 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 zonegroup default 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.886 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.885+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.887 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.886+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.887 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.886+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.887 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.886+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.887 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.886+0000 7f0a977a3900 20 started sync module instance, tier type = 2026-04-17T12:20:43.887 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.886+0000 7f0a977a3900 20 started zone id=f9f3cc56-b810-479f-9b9d-92717e61b00d (name=default) with tier type = 2026-04-17T12:20:43.894 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.893+0000 7f0a977a3900 20 add_watcher() i=7 2026-04-17T12:20:43.895 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.894+0000 7f0a977a3900 20 add_watcher() i=3 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 20 add_watcher() i=5 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 20 add_watcher() i=6 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 20 add_watcher() i=0 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 20 add_watcher() i=4 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 20 add_watcher() i=2 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 20 add_watcher() i=1 2026-04-17T12:20:43.896 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.895+0000 7f0a977a3900 2 all 8 watchers are set, enabling cache 2026-04-17T12:20:43.898 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.897+0000 7f0a977a3900 20 rgw_check_secure_mon_conn(): auth registy supported: methods=[2] modes=[2,1] 2026-04-17T12:20:43.898 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.897+0000 7f0a977a3900 20 rgw_check_secure_mon_conn(): mode 1 is insecure 2026-04-17T12:20:43.898 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.897+0000 7f0a977a3900 5 note: GC not initialized 2026-04-17T12:20:43.898 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.898+0000 7f0a3bfff640 20 reqs_thread_entry: start 2026-04-17T12:20:43.942 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.941+0000 7f0a977a3900 20 init_complete bucket index max shards: 11 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.941+0000 7f0a977a3900 20 Filter name: none 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.941+0000 7f0a39ffb640 20 reqs_thread_entry: start 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.941+0000 7f0a977a3900 10 cache get: name=default.rgw.meta+users.uid+foo.client.0 : miss 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.941+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 cache put: name=default.rgw.meta+users.uid+foo.client.0 info.flags=0x0 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 adding default.rgw.meta+users.uid+foo.client.0 to cache LRU end 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 cache get: name=default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU : miss 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 20 rados->read ofs=0 len=0 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 20 rados_obj.operate() r=-2 bl.length=0 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 cache put: name=default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU info.flags=0x0 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 adding default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU to cache LRU end 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 cache get: name=default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU : hit (negative entry) 2026-04-17T12:20:43.943 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.942+0000 7f0a977a3900 10 cache get: name=default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU : hit (negative entry) 2026-04-17T12:20:43.944 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.943+0000 7f0a977a3900 10 cache put: name=default.rgw.meta+users.uid+foo.client.0 info.flags=0x17 2026-04-17T12:20:43.944 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.943+0000 7f0a977a3900 10 moving default.rgw.meta+users.uid+foo.client.0 to cache LRU end 2026-04-17T12:20:43.944 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.943+0000 7f0a977a3900 10 distributing notification oid=default.rgw.control:notify.0 cni=[op: 0, obj: default.rgw.meta:users.uid:foo.client.0, ofs0, ns] 2026-04-17T12:20:43.945 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.944+0000 7f0a6e7fc640 10 rgw watcher librados: RGWWatcher::handle_notify() notify_id 163208757248 cookie 93868168057040 notifier 4688 bl.length()=628 2026-04-17T12:20:43.945 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.944+0000 7f0a6e7fc640 10 rgw watcher librados: cache put: name=default.rgw.meta+users.uid+foo.client.0 info.flags=0x17 2026-04-17T12:20:43.945 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.944+0000 7f0a6e7fc640 10 rgw watcher librados: moving default.rgw.meta+users.uid+foo.client.0 to cache LRU end 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.946+0000 7f0a977a3900 10 cache put: name=default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU info.flags=0x7 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.946+0000 7f0a977a3900 10 moving default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU to cache LRU end 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.946+0000 7f0a977a3900 10 distributing notification oid=default.rgw.control:notify.2 cni=[op: 0, obj: default.rgw.meta:users.keys:WZNRJECYHIJXXDDQHDUU, ofs0, ns] 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.946+0000 7f0a6e7fc640 10 rgw watcher librados: RGWWatcher::handle_notify() notify_id 163208757248 cookie 93868169643520 notifier 4688 bl.length()=186 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.946+0000 7f0a6e7fc640 10 rgw watcher librados: cache put: name=default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU info.flags=0x7 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.946+0000 7f0a6e7fc640 10 rgw watcher librados: moving default.rgw.meta+users.keys+WZNRJECYHIJXXDDQHDUU to cache LRU end 2026-04-17T12:20:43.947 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "user_id": "foo.client.0", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "display_name": "Mr. foo.client.0", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "email": "", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "suspended": 0, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_buckets": 1000, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "subusers": [], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "keys": [ 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: { 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "user": "foo.client.0", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "access_key": "WZNRJECYHIJXXDDQHDUU", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "secret_key": "LVqqWb6ixrnFJANAuw7aL+nM3sZlqbJRogGtsy3lZjDz6pzRhxNPXw==", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "active": true, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "create_date": "2026-04-17T12:20:43.943335Z" 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: } 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: ], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "swift_keys": [], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "caps": [], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "op_mask": "read, write, delete", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "default_placement": "", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "default_storage_class": "", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "placement_tags": [], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "bucket_quota": { 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "enabled": false, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "check_on_raw": false, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_size": -1, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_size_kb": 0, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_objects": -1 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: }, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "user_quota": { 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "enabled": false, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "check_on_raw": false, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_size": -1, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_size_kb": 0, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "max_objects": -1 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: }, 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "temp_url_keys": [], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "type": "rgw", 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "mfa_ids": [], 2026-04-17T12:20:43.948 INFO:teuthology.orchestra.run.vm02.stdout: "account_id": "", 2026-04-17T12:20:43.949 INFO:teuthology.orchestra.run.vm02.stdout: "path": "/", 2026-04-17T12:20:43.949 INFO:teuthology.orchestra.run.vm02.stdout: "create_date": "2026-04-17T12:20:43.943329Z", 2026-04-17T12:20:43.949 INFO:teuthology.orchestra.run.vm02.stdout: "tags": [], 2026-04-17T12:20:43.949 INFO:teuthology.orchestra.run.vm02.stdout: "group_ids": [] 2026-04-17T12:20:43.949 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-04-17T12:20:43.949 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:20:43.952 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.951+0000 7f0a977a3900 20 remove_watcher() i=1 2026-04-17T12:20:43.952 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.951+0000 7f0a977a3900 2 removed watcher, disabling cache 2026-04-17T12:20:43.952 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.951+0000 7f0a977a3900 20 remove_watcher() i=0 2026-04-17T12:20:43.952 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.951+0000 7f0a977a3900 20 remove_watcher() i=3 2026-04-17T12:20:43.952 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.952+0000 7f0a977a3900 20 remove_watcher() i=2 2026-04-17T12:20:43.953 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.952+0000 7f0a977a3900 20 remove_watcher() i=6 2026-04-17T12:20:43.953 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.952+0000 7f0a977a3900 20 remove_watcher() i=5 2026-04-17T12:20:43.953 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.952+0000 7f0a977a3900 20 remove_watcher() i=4 2026-04-17T12:20:43.953 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:20:43.952+0000 7f0a977a3900 20 remove_watcher() i=7 2026-04-17T12:20:43.959 INFO:tasks.dedup_tests:Configuring dedup-tests... 2026-04-17T12:20:43.960 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-04-17T12:20:43.960 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/deduptests.client.0.conf 2026-04-17T12:20:43.978 INFO:tasks.dedup_tests:Running dedup-tests... 2026-04-17T12:20:43.978 DEBUG:teuthology.orchestra.run.vm02:dedup tests against rgw> source /home/ubuntu/cephtest/tox-venv/bin/activate && cd /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/ && DEDUPTESTS_CONF=./deduptests.client.0.conf tox -- -v -m 'basic_test or request_test or example_test' 2026-04-17T12:20:44.486 INFO:teuthology.orchestra.run.vm02.stdout:py: install_deps> python -I -m pip install -r requirements.txt 2026-04-17T12:20:47.313 INFO:teuthology.orchestra.run.vm02.stdout:py: commands[0]> pytest -v -m 'basic_test or request_test or example_test' 2026-04-17T12:20:47.400 INFO:teuthology.orchestra.run.vm02.stdout:============================= test session starts ============================== 2026-04-17T12:20:47.400 INFO:teuthology.orchestra.run.vm02.stdout:platform linux -- Python 3.9.23, pytest-8.4.2, pluggy-1.6.0 -- /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/bin/python 2026-04-17T12:20:47.400 INFO:teuthology.orchestra.run.vm02.stdout:cachedir: .tox/py/.pytest_cache 2026-04-17T12:20:47.400 INFO:teuthology.orchestra.run.vm02.stdout:rootdir: /home/ubuntu/cephtest/ceph/src/test/rgw/dedup 2026-04-17T12:20:47.400 INFO:teuthology.orchestra.run.vm02.stdout:configfile: pytest.ini 2026-04-17T12:20:47.494 INFO:teuthology.orchestra.run.vm02.stdout:collecting ... collected 34 items 2026-04-17T12:20:47.494 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:20:47.624 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_etag_corruption PASSED [ 2%] 2026-04-17T12:20:47.624 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_md5_collisions PASSED [ 5%] 2026-04-17T12:20:47.624 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_small PASSED [ 8%] 2026-04-17T12:20:47.625 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_small_with_tenants PASSED [ 11%] 2026-04-17T12:20:47.625 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_0_with_tenants PASSED [ 14%] 2026-04-17T12:20:47.625 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_0 PASSED [ 17%] 2026-04-17T12:20:47.626 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_1_with_tenants PASSED [ 20%] 2026-04-17T12:20:47.626 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_1 PASSED [ 23%] 2026-04-17T12:20:47.626 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_2_with_tenants PASSED [ 26%] 2026-04-17T12:20:47.627 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_2 PASSED [ 29%] 2026-04-17T12:20:47.627 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_with_remove_multi_tenants PASSED [ 32%] 2026-04-17T12:20:47.627 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_with_remove PASSED [ 35%] 2026-04-17T12:20:47.628 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_multipart_with_tenants PASSED [ 38%] 2026-04-17T12:20:47.628 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_multipart PASSED [ 41%] 2026-04-17T12:20:47.628 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_basic_with_tenants PASSED [ 44%] 2026-04-17T12:20:47.629 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_basic PASSED [ 47%] 2026-04-17T12:20:47.629 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_small_multipart_with_tenants PASSED [ 50%] 2026-04-17T12:20:47.629 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_small_multipart PASSED [ 52%] 2026-04-17T12:20:47.630 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_large_scale_with_tenants PASSED [ 55%] 2026-04-17T12:20:47.631 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_large_scale PASSED [ 58%] 2026-04-17T12:20:47.631 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_empty_bucket PASSED [ 61%] 2026-04-17T12:20:47.632 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_inc_loop_with_tenants PASSED [ 64%] 2026-04-17T12:20:54.114 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_with_tenants 2026-04-17T12:20:54.114 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:20:54.114 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:20:54.681 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 67%] 2026-04-17T12:23:16.626 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_multipart 2026-04-17T12:23:16.626 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:23:16.626 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:23:20.670 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 70%] 2026-04-17T12:23:30.058 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_basic 2026-04-17T12:23:30.058 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:23:30.058 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:23:30.524 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 73%] 2026-04-17T12:23:40.399 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_multipart 2026-04-17T12:23:40.399 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:23:40.399 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:23:40.866 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 76%] 2026-04-17T12:23:46.664 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small 2026-04-17T12:23:46.664 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:23:46.664 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:23:47.091 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 79%] 2026-04-17T12:24:01.484 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_large_mix 2026-04-17T12:24:01.484 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:24:01.484 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:24:02.667 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 82%] 2026-04-17T12:24:23.118 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_basic_with_tenants 2026-04-17T12:24:23.118 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:24:23.118 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:24:24.140 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 85%] 2026-04-17T12:25:24.289 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_multipart_with_tenants 2026-04-17T12:25:24.289 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:25:24.289 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:25:26.256 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 88%] 2026-04-17T12:25:35.657 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_multipart_with_tenants 2026-04-17T12:25:35.657 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:25:35.657 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:25:36.256 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 91%] 2026-04-17T12:30:22.726 INFO:tasks.ceph.mon.a.vm02.stderr:2026-04-17T12:30:22.725+0000 7f268a4fe640 -1 log_channel(cluster) log [ERR] : Health check failed: mon c is very low on available space (MON_DISK_CRIT) 2026-04-17T12:30:29.914 INFO:tasks.ceph.mon.a.vm02.stderr:2026-04-17T12:30:29.913+0000 7f268cd03640 -1 log_channel(cluster) log [ERR] : Health check update: mons a,c are very low on available space (MON_DISK_CRIT) 2026-04-17T12:32:51.834 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_large_scale_with_tenants 2026-04-17T12:32:51.835 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:32:51.835 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1096 dedup completed in 5 seconds 2026-04-17T12:32:51.835 INFO:teuthology.orchestra.run.vm02.stdout:INFO dedup.test_dedup:test_dedup.py:1288 [64] obj_count=65526, upload=419(sec), exec=5(sec), verify=0(sec) 2026-04-17T12:33:04.936 INFO:tasks.ceph.mon.a.vm02.stderr:2026-04-17T12:33:04.935+0000 7f268cd03640 -1 log_channel(cluster) log [ERR] : Health check update: mons a,b,c are very low on available space (MON_DISK_CRIT) 2026-04-17T12:35:03.035 INFO:teuthology.orchestra.run.vm02.stdout:PASSED [ 94%] 2026-04-17T12:35:06.510 INFO:tasks.rgw.client.1.vm06.stdout:2026-04-17T12:35:06.509+0000 7f2a23039640 -1 restore: virtual void* rgw::restore::Restore::RestoreWorker::entry(): ERROR: restore process() returned error r=-16 2026-04-17T12:35:36.551 INFO:tasks.ceph.osd.0.vm02.stderr:problem writing to /var/log/ceph/ceph-osd.0.log: (28) No space left on device 2026-04-17T12:35:36.551 INFO:tasks.ceph.osd.1.vm02.stderr:problem writing to /var/log/ceph/ceph-osd.1.log: (28) No space left on device 2026-04-17T12:35:36.552 INFO:tasks.ceph.osd.2.vm02.stderr:problem writing to /var/log/ceph/ceph-osd.2.log: (28) No space left on device 2026-04-17T12:35:36.552 INFO:tasks.ceph.osd.3.vm02.stderr:problem writing to /var/log/ceph/ceph-osd.3.log: (28) No space left on device 2026-04-17T12:35:36.552 INFO:tasks.ceph.osd.1.vm02.stderr:problem writing to /var/log/ceph/ceph-osd.1.log: (28) No space left on device 2026-04-17T12:35:36.553 INFO:tasks.rgw.client.0.vm02.stdout:problem writing to /var/log/ceph/rgw.ceph.client.0.log: (28) No space left on device 2026-04-17T12:35:36.553 INFO:tasks.rgw.client.0.vm02.stdout:tee: /var/log/ceph/rgw.ceph.client.0.stdout: No space left on device 2026-04-17T12:35:36.691 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:35:36.777 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:35:40.079 INFO:tasks.ceph.mgr.y.vm02.stderr:problem writing to /var/log/ceph/ceph-mgr.y.log: (28) No space left on device 2026-04-17T12:36:28.369 INFO:tasks.ceph.osd.4.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.4.log: (28) No space left on device 2026-04-17T12:36:28.371 INFO:tasks.ceph.osd.5.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.5.log: (28) No space left on device 2026-04-17T12:36:28.371 INFO:tasks.ceph.osd.6.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.6.log: (28) No space left on device 2026-04-17T12:36:28.371 INFO:tasks.ceph.osd.7.vm06.stderr:problem writing to /var/log/ceph/ceph-osd.7.log: (28) No space left on device 2026-04-17T12:36:28.378 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:28.820 INFO:tasks.ceph.mgr.x.vm06.stderr:problem writing to /var/log/ceph/ceph-mgr.x.log: (28) No space left on device 2026-04-17T12:36:36.415 INFO:tasks.ceph.mon.a.vm02.stderr:2026-04-17T12:36:36.414+0000 7f268cd03640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-a/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:36.415 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = '1165' value size = 611) 2026-04-17T12:36:36.415 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = 'pending_v' value size = 8) 2026-04-17T12:36:36.415 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = 'pending_pn' value size = 8) 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7f268cd03640 time 2026-04-17T12:36:36.417362+0000 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7f2692b901fd] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 4: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 6: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 7: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr:2026-04-17T12:36:36.426+0000 7f268cd03640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7f268cd03640 time 2026-04-17T12:36:36.417362+0000 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.426 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7f2692b901fd] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 4: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 6: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 7: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.427 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.428 INFO:tasks.ceph.mon.a.vm02.stderr:*** Caught signal (Aborted) ** 2026-04-17T12:36:36.428 INFO:tasks.ceph.mon.a.vm02.stderr: in thread 7f268cd03640 thread_name:safe_timer 2026-04-17T12:36:36.430 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.430 INFO:tasks.ceph.mon.a.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7f2691c3fc30] 2026-04-17T12:36:36.430 INFO:tasks.ceph.mon.a.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7f2691c8d02c] 2026-04-17T12:36:36.430 INFO:tasks.ceph.mon.a.vm02.stderr: 3: raise() 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 4: abort() 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7f2692b902ba] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 9: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 10: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 11: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr:2026-04-17T12:36:36.430+0000 7f268cd03640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: in thread 7f268cd03640 thread_name:safe_timer 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7f2691c3fc30] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7f2691c8d02c] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 3: raise() 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 4: abort() 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7f2692b902ba] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 9: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 10: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 11: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:36.431 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.432 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: -2> 2026-04-17T12:36:36.414+0000 7f268cd03640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-a/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = '1165' value size = 611) 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = 'pending_v' value size = 8) 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = 'pending_pn' value size = 8) 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: -1> 2026-04-17T12:36:36.426+0000 7f268cd03640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7f268cd03640 time 2026-04-17T12:36:36.417362+0000 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7f2692b901fd] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 4: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 6: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 7: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 0> 2026-04-17T12:36:36.430+0000 7f268cd03640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: in thread 7f268cd03640 thread_name:safe_timer 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7f2691c3fc30] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7f2691c8d02c] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 3: raise() 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 4: abort() 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7f2692b902ba] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 9: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 10: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 11: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:36.446 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.449 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.450 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.a.log: (28) No space left on device 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: -9999> 2026-04-17T12:36:36.414+0000 7f268cd03640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-a/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = '1165' value size = 611) 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = 'pending_v' value size = 8) 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:PutCF( prefix = paxos key = 'pending_pn' value size = 8) 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: -9998> 2026-04-17T12:36:36.426+0000 7f268cd03640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7f268cd03640 time 2026-04-17T12:36:36.417362+0000 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7f2692b901fd] 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.451 INFO:tasks.ceph.mon.a.vm02.stderr: 3: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 4: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 6: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 7: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 9: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 10: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 11: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: -9997> 2026-04-17T12:36:36.430+0000 7f268cd03640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: in thread 7f268cd03640 thread_name:safe_timer 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7f2691c3fc30] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7f2691c8d02c] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 3: raise() 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 4: abort() 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7f2692b902ba] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x564bba8a01ac] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 7: (Paxos::begin(ceph::buffer::v15_2_0::list&)+0x54c) [0x564bbaa1e95c] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 8: (Paxos::propose_pending()+0x11b) [0x564bbaa2c70b] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 9: (Paxos::trigger_propose()+0x118) [0x564bbaa2cb08] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 10: (PaxosService::propose_pending()+0x176) [0x564bbaa2ce46] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 11: ceph-mon(+0x2a644d) [0x564bba8a044d] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 12: (CommonSafeTimer::timer_thread()+0x130) [0x7f2692cdc550] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 13: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7f2692cdcfb1] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 14: /lib64/libc.so.6(+0x8b2ea) [0x7f2691c8b2ea] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 15: /lib64/libc.so.6(+0x1103c0) [0x7f2691d103c0] 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:36.452 INFO:tasks.ceph.mon.a.vm02.stderr: 2026-04-17T12:36:36.521 INFO:tasks.ceph.mon.a.vm02.stderr:daemon-helper: command crashed with signal 6 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr:2026-04-17T12:36:39.916+0000 7efd39ca0640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-c/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7efd39ca0640 time 2026-04-17T12:36:39.917823+0000 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7efd423901fd] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 9: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr:2026-04-17T12:36:39.918+0000 7efd39ca0640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7efd39ca0640 time 2026-04-17T12:36:39.917823+0000 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7efd423901fd] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 9: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.918 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr:*** Caught signal (Aborted) ** 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: in thread 7efd39ca0640 thread_name:ms_dispatch 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7efd4143fc30] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7efd4148d02c] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 3: raise() 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 4: abort() 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7efd423902ba] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 13: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr:2026-04-17T12:36:39.918+0000 7efd39ca0640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: in thread 7efd39ca0640 thread_name:ms_dispatch 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7efd4143fc30] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7efd4148d02c] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 3: raise() 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 4: abort() 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7efd423902ba] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 13: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:39.919 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.920 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.933 INFO:tasks.ceph.mon.c.vm02.stderr: -2> 2026-04-17T12:36:39.916+0000 7efd39ca0640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-c/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: -1> 2026-04-17T12:36:39.918+0000 7efd39ca0640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7efd39ca0640 time 2026-04-17T12:36:39.917823+0000 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7efd423901fd] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 9: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 0> 2026-04-17T12:36:39.918+0000 7efd39ca0640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: in thread 7efd39ca0640 thread_name:ms_dispatch 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7efd4143fc30] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7efd4148d02c] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 3: raise() 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 4: abort() 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7efd423902ba] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.934 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 13: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:39.935 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.936 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.937 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.938 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.939 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:problem writing to /var/log/ceph/ceph-mon.c.log: (28) No space left on device 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr: -9999> 2026-04-17T12:36:39.916+0000 7efd39ca0640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-c/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr: -9998> 2026-04-17T12:36:39.918+0000 7efd39ca0640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7efd39ca0640 time 2026-04-17T12:36:39.917823+0000 2026-04-17T12:36:39.940 INFO:tasks.ceph.mon.c.vm02.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7efd423901fd] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 2: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 4: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 6: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 9: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 11: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 12: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 13: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: -9997> 2026-04-17T12:36:39.918+0000 7efd39ca0640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: in thread 7efd39ca0640 thread_name:ms_dispatch 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7efd4143fc30] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7efd4148d02c] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 3: raise() 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 4: abort() 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7efd423902ba] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 6: ceph-mon(+0x2a61ac) [0x560caf96d1ac] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x560cafa50865] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 8: (ConnectionTracker::report_live_connection(int, double)+0x181) [0x560cafa59901] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 9: (Elector::handle_ping(boost::intrusive_ptr)+0x620) [0x560cafa55cd0] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 10: (Elector::dispatch(boost::intrusive_ptr)+0xa7) [0x560cafa56887] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 11: (Monitor::dispatch_op(boost::intrusive_ptr)+0xe4d) [0x560caf9c5e3d] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 12: (Monitor::_ms_dispatch(Message*)+0x786) [0x560caf9ba5c6] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 13: ceph-mon(+0x2b333c) [0x560caf97a33c] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 14: (DispatchQueue::entry()+0x4a8) [0x7efd42606848] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 15: /usr/lib64/ceph/libceph-common.so.2(+0x49ac51) [0x7efd4269ac51] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 16: /lib64/libc.so.6(+0x8b2ea) [0x7efd4148b2ea] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 17: /lib64/libc.so.6(+0x1103c0) [0x7efd415103c0] 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:39.941 INFO:tasks.ceph.mon.c.vm02.stderr: 2026-04-17T12:36:40.140 INFO:tasks.ceph.mon.c.vm02.stderr:daemon-helper: command crashed with signal 6 2026-04-17T12:36:40.373 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~0s 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr:2026-04-17T12:36:41.400+0000 7fb571aa9640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-b/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fb571aa9640 time 2026-04-17T12:36:41.401862+0000 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fb5779901fd] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.401 INFO:tasks.ceph.mon.b.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr:2026-04-17T12:36:41.401+0000 7fb571aa9640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fb571aa9640 time 2026-04-17T12:36:41.401862+0000 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fb5779901fd] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr:*** Caught signal (Aborted) ** 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: in thread 7fb571aa9640 thread_name:safe_timer 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fb576a3fc30] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fb576a8d02c] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 3: raise() 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 4: abort() 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fb5779902ba] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 9: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.402 INFO:tasks.ceph.mon.b.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr:2026-04-17T12:36:41.402+0000 7fb571aa9640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: in thread 7fb571aa9640 thread_name:safe_timer 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fb576a3fc30] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fb576a8d02c] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 3: raise() 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 4: abort() 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fb5779902ba] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 9: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.403 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: -2> 2026-04-17T12:36:41.400+0000 7fb571aa9640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-b/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: -1> 2026-04-17T12:36:41.401+0000 7fb571aa9640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fb571aa9640 time 2026-04-17T12:36:41.401862+0000 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fb5779901fd] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.414 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 0> 2026-04-17T12:36:41.402+0000 7fb571aa9640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: in thread 7fb571aa9640 thread_name:safe_timer 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fb576a3fc30] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fb576a8d02c] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 3: raise() 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 4: abort() 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fb5779902ba] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 9: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.415 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.418 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.419 INFO:tasks.ceph.mon.b.vm06.stderr:problem writing to /var/log/ceph/ceph-mon.b.log: (28) No space left on device 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: -9999> 2026-04-17T12:36:41.400+0000 7fb571aa9640 -1 rocksdb: submit_common error: IO error: No space left on device: While open a file for appending: /var/lib/ceph/mon/ceph-b/store.db/000022.log: No space left on device code =  Rocksdb transaction: 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr:PutCF( prefix = monitor key = 'connectivity_scores' value size = 238) 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: -9998> 2026-04-17T12:36:41.401+0000 7fb571aa9640 -1 /runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: In function 'int MonitorDBStore::apply_transaction(TransactionRef)' thread 7fb571aa9640 time 2026-04-17T12:36:41.401862+0000 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr:/runner/scratch/rpms/ceph-debug/20.2.0-21-gc03ba9ecf58/BUILD/ceph-20.2.0-21-gc03ba9ecf58/src/mon/MonitorDBStore.h: 356: ceph_abort_msg("failed to write to db") 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 1: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0xc9) [0x7fb5779901fd] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 2: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 3: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 4: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 8: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 9: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 10: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: -9997> 2026-04-17T12:36:41.402+0000 7fb571aa9640 -1 *** Caught signal (Aborted) ** 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: in thread 7fb571aa9640 thread_name:safe_timer 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: ceph version 20.2.0-21-gc03ba9ecf58 (c03ba9ecf58a4116bdd5049c6e392c7a287bc4f8) tentacle (stable - RelWithDebInfo) 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 1: /lib64/libc.so.6(+0x3fc30) [0x7fb576a3fc30] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 2: /lib64/libc.so.6(+0x8d02c) [0x7fb576a8d02c] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 3: raise() 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 4: abort() 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 5: (ceph::__ceph_abort(char const*, int, char const*, std::__cxx11::basic_string, std::allocator > const&)+0x186) [0x7fb5779902ba] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 6: ceph-mon(+0x2a61ac) [0x55c9a15e71ac] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 7: (Elector::persist_connectivity_scores()+0x135) [0x55c9a16ca865] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 8: (ConnectionTracker::report_dead_connection(int, double)+0x181) [0x55c9a16d40c1] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 9: (Elector::ping_check(int)+0x2e2) [0x55c9a16cf182] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 10: ceph-mon(+0x2a644d) [0x55c9a15e744d] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 11: (CommonSafeTimer::timer_thread()+0x130) [0x7fb577adc550] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 12: /usr/lib64/ceph/libceph-common.so.2(+0x2dcfb1) [0x7fb577adcfb1] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 13: /lib64/libc.so.6(+0x8b2ea) [0x7fb576a8b2ea] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 14: /lib64/libc.so.6(+0x1103c0) [0x7fb576b103c0] 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this. 2026-04-17T12:36:41.420 INFO:tasks.ceph.mon.b.vm06.stderr: 2026-04-17T12:36:41.426 INFO:tasks.ceph.mon.b.vm06.stderr:daemon-helper: command crashed with signal 6 2026-04-17T12:36:46.681 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~7s 2026-04-17T12:36:46.681 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~0s 2026-04-17T12:36:46.681 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~0s 2026-04-17T12:36:52.989 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~13s 2026-04-17T12:36:52.989 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~6s 2026-04-17T12:36:52.989 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~6s 2026-04-17T12:36:59.299 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~19s 2026-04-17T12:36:59.299 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~13s 2026-04-17T12:36:59.299 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~13s 2026-04-17T12:37:05.609 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~25s 2026-04-17T12:37:05.609 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~19s 2026-04-17T12:37:05.609 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~19s 2026-04-17T12:37:11.917 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~32s 2026-04-17T12:37:11.917 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~25s 2026-04-17T12:37:11.917 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~25s 2026-04-17T12:37:18.225 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~38s 2026-04-17T12:37:18.226 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~32s 2026-04-17T12:37:18.226 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~32s 2026-04-17T12:37:24.533 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~44s 2026-04-17T12:37:24.533 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~38s 2026-04-17T12:37:24.533 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~38s 2026-04-17T12:37:30.840 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~51s 2026-04-17T12:37:30.840 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~44s 2026-04-17T12:37:30.840 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~44s 2026-04-17T12:37:37.146 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~57s 2026-04-17T12:37:37.146 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~50s 2026-04-17T12:37:37.146 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~50s 2026-04-17T12:37:43.452 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~63s 2026-04-17T12:37:43.453 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~57s 2026-04-17T12:37:43.453 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~57s 2026-04-17T12:37:49.759 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~70s 2026-04-17T12:37:49.759 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~63s 2026-04-17T12:37:49.759 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~63s 2026-04-17T12:37:51.407 INFO:tasks.rgw.client.1.vm06.stdout:problem writing to /var/log/ceph/rgw.ceph.client.1.log: (28) No space left on device 2026-04-17T12:37:56.066 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~76s 2026-04-17T12:37:56.066 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~69s 2026-04-17T12:37:56.066 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~69s 2026-04-17T12:38:02.372 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~82s 2026-04-17T12:38:02.372 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~76s 2026-04-17T12:38:02.372 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~76s 2026-04-17T12:38:08.681 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~89s 2026-04-17T12:38:08.681 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~82s 2026-04-17T12:38:08.681 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~82s 2026-04-17T12:38:14.989 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~95s 2026-04-17T12:38:14.989 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~88s 2026-04-17T12:38:14.989 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~88s 2026-04-17T12:38:21.298 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~101s 2026-04-17T12:38:21.298 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~95s 2026-04-17T12:38:21.298 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~95s 2026-04-17T12:38:27.605 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~107s 2026-04-17T12:38:27.605 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~101s 2026-04-17T12:38:27.605 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~101s 2026-04-17T12:38:33.912 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~114s 2026-04-17T12:38:33.912 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~107s 2026-04-17T12:38:33.912 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~107s 2026-04-17T12:38:40.219 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~120s 2026-04-17T12:38:40.219 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~114s 2026-04-17T12:38:40.219 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~114s 2026-04-17T12:38:46.525 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~126s 2026-04-17T12:38:46.525 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~120s 2026-04-17T12:38:46.525 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~120s 2026-04-17T12:38:52.832 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~133s 2026-04-17T12:38:52.832 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~126s 2026-04-17T12:38:52.832 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~126s 2026-04-17T12:38:59.139 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~139s 2026-04-17T12:38:59.139 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~132s 2026-04-17T12:38:59.139 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~132s 2026-04-17T12:39:05.446 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~145s 2026-04-17T12:39:05.446 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~139s 2026-04-17T12:39:05.446 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~139s 2026-04-17T12:39:11.752 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~152s 2026-04-17T12:39:11.752 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~145s 2026-04-17T12:39:11.752 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~145s 2026-04-17T12:39:18.058 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~158s 2026-04-17T12:39:18.058 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~151s 2026-04-17T12:39:18.058 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~151s 2026-04-17T12:39:24.365 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~164s 2026-04-17T12:39:24.365 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~158s 2026-04-17T12:39:24.365 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~158s 2026-04-17T12:39:30.676 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~171s 2026-04-17T12:39:30.676 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~164s 2026-04-17T12:39:30.676 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~164s 2026-04-17T12:39:36.988 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~177s 2026-04-17T12:39:36.988 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~170s 2026-04-17T12:39:36.988 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~170s 2026-04-17T12:39:43.297 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~183s 2026-04-17T12:39:43.297 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~177s 2026-04-17T12:39:43.297 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~177s 2026-04-17T12:39:49.602 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~189s 2026-04-17T12:39:49.602 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~183s 2026-04-17T12:39:49.602 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~183s 2026-04-17T12:39:55.908 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~196s 2026-04-17T12:39:55.908 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~189s 2026-04-17T12:39:55.908 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~189s 2026-04-17T12:40:02.213 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~202s 2026-04-17T12:40:02.214 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~196s 2026-04-17T12:40:02.214 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~196s 2026-04-17T12:40:08.520 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~208s 2026-04-17T12:40:08.521 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~202s 2026-04-17T12:40:08.521 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~202s 2026-04-17T12:40:14.828 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~215s 2026-04-17T12:40:14.828 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~208s 2026-04-17T12:40:14.828 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~208s 2026-04-17T12:40:21.134 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~221s 2026-04-17T12:40:21.134 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~214s 2026-04-17T12:40:21.134 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~214s 2026-04-17T12:40:27.440 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~227s 2026-04-17T12:40:27.440 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~221s 2026-04-17T12:40:27.440 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~221s 2026-04-17T12:40:33.746 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~234s 2026-04-17T12:40:33.746 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~227s 2026-04-17T12:40:33.746 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~227s 2026-04-17T12:40:40.052 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~240s 2026-04-17T12:40:40.052 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~233s 2026-04-17T12:40:40.052 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~233s 2026-04-17T12:40:46.358 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~246s 2026-04-17T12:40:46.358 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~240s 2026-04-17T12:40:46.358 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~240s 2026-04-17T12:40:52.664 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~252s 2026-04-17T12:40:52.665 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~246s 2026-04-17T12:40:52.665 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~246s 2026-04-17T12:40:58.972 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~259s 2026-04-17T12:40:58.972 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~252s 2026-04-17T12:40:58.972 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~252s 2026-04-17T12:41:05.278 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~265s 2026-04-17T12:41:05.278 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~259s 2026-04-17T12:41:05.278 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~259s 2026-04-17T12:41:11.584 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~271s 2026-04-17T12:41:11.585 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~265s 2026-04-17T12:41:11.585 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~265s 2026-04-17T12:41:17.891 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~278s 2026-04-17T12:41:17.891 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~271s 2026-04-17T12:41:17.891 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~271s 2026-04-17T12:41:24.197 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~284s 2026-04-17T12:41:24.197 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~278s 2026-04-17T12:41:24.197 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~278s 2026-04-17T12:41:30.503 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~290s 2026-04-17T12:41:30.503 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~284s 2026-04-17T12:41:30.503 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~284s 2026-04-17T12:41:36.809 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~297s 2026-04-17T12:41:36.809 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~290s 2026-04-17T12:41:36.809 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~290s 2026-04-17T12:41:43.116 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.a is failed for ~303s 2026-04-17T12:41:43.117 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.c is failed for ~296s 2026-04-17T12:41:43.117 INFO:tasks.daemonwatchdog.daemon_watchdog:daemon ceph.mon.b is failed for ~296s 2026-04-17T12:41:43.117 INFO:tasks.daemonwatchdog.daemon_watchdog:BARK! unmounting mounts and killing all daemons 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.0:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.1:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.2:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.3:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.4:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.5:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.6:Sent signal 15 2026-04-17T12:41:44.427 INFO:tasks.ceph.osd.7:Sent signal 15 2026-04-17T12:41:44.428 INFO:tasks.rgw.client.0:Sent signal 15 2026-04-17T12:41:44.428 INFO:tasks.rgw.client.1:Sent signal 15 2026-04-17T12:41:44.428 INFO:tasks.rgw.client.2:Sent signal 15 2026-04-17T12:41:44.428 INFO:tasks.ceph.mgr.y:Sent signal 15 2026-04-17T12:41:44.428 INFO:tasks.ceph.mgr.x:Sent signal 15 2026-04-17T12:41:44.428 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:41:44.428+0000 7f90388dd640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 4 (PID: 59800) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:41:44.428+0000 7f90388dd640 -1 osd.4 71 *** Got signal Terminated *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.4.vm06.stderr:2026-04-17T12:41:44.428+0000 7f90388dd640 -1 osd.4 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:41:44.428+0000 7fd8972e2640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 7 (PID: 59801) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:41:44.428+0000 7fd8972e2640 -1 osd.7 71 *** Got signal Terminated *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.7.vm06.stderr:2026-04-17T12:41:44.428+0000 7fd8972e2640 -1 osd.7 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:41:44.428+0000 7fe101328640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 0 (PID: 60489) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:41:44.428+0000 7fe101328640 -1 osd.0 71 *** Got signal Terminated *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:41:44.428+0000 7f51859bc640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 2 (PID: 60495) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:41:44.428+0000 7f51859bc640 -1 osd.2 71 *** Got signal Terminated *** 2026-04-17T12:41:44.429 INFO:tasks.rgw.client.1.vm06.stdout:2026-04-17T12:41:44.428+0000 7f2b48a84640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper term radosgw --rgw-frontends beast port=80 -n client.1 --cluster ceph -k /etc/ceph/ceph.client.1.keyring --log-file /var/log/ceph/rgw.ceph.client.1.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.1.sock --foreground (PID: 64101) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.rgw.client.1.vm06.stdout:2026-04-17T12:41:44.428+0000 7f2b4e12a980 -1 shutting down 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:41:44.428+0000 7f217f188640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 3 (PID: 60502) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:41:44.428+0000 7f217f188640 -1 osd.3 71 *** Got signal Terminated *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:41:44.429+0000 7f583ef9c640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 1 (PID: 60488) UID: 0 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:41:44.429+0000 7f583ef9c640 -1 osd.1 71 *** Got signal Terminated *** 2026-04-17T12:41:44.429 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:41:44.429+0000 7f729b1e4640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 6 (PID: 59798) UID: 0 2026-04-17T12:41:44.430 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:41:44.429+0000 7f729b1e4640 -1 osd.6 71 *** Got signal Terminated *** 2026-04-17T12:41:44.430 INFO:tasks.ceph.osd.6.vm06.stderr:2026-04-17T12:41:44.429+0000 7f729b1e4640 -1 osd.6 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.430 INFO:tasks.rgw.client.0.vm02.stdout:2026-04-17T12:41:44.429+0000 7f42740ac640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper term radosgw --rgw-frontends beast port=80 -n client.0 --cluster ceph -k /etc/ceph/ceph.client.0.keyring --log-file /var/log/ceph/rgw.ceph.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock --foreground (PID: 66553) UID: 0 2026-04-17T12:41:44.430 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:41:44.429+0000 7f9eeb4c2640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper kill ceph-osd -f --cluster ceph -i 5 (PID: 59793) UID: 0 2026-04-17T12:41:44.430 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:41:44.429+0000 7f9eeb4c2640 -1 osd.5 71 *** Got signal Terminated *** 2026-04-17T12:41:44.430 INFO:tasks.ceph.osd.5.vm06.stderr:2026-04-17T12:41:44.429+0000 7f9eeb4c2640 -1 osd.5 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.430 INFO:tasks.rgw.client.2.vm08.stdout:2026-04-17T12:41:44.429+0000 7fe7e010a640 -1 received signal: Terminated from /usr/bin/python3 /bin/daemon-helper term radosgw --rgw-frontends beast port=80 -n client.2 --cluster ceph -k /etc/ceph/ceph.client.2.keyring --log-file /var/log/ceph/rgw.ceph.client.2.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.2.sock --foreground (PID: 53232) UID: 0 2026-04-17T12:41:44.430 INFO:tasks.rgw.client.2.vm08.stdout:2026-04-17T12:41:44.429+0000 7fe7e5923980 -1 shutting down 2026-04-17T12:41:44.433 INFO:tasks.rgw.client.0.vm02.stdout:2026-04-17T12:41:44.433+0000 7f427790f980 -1 shutting down 2026-04-17T12:41:44.434 INFO:tasks.ceph.osd.2.vm02.stderr:2026-04-17T12:41:44.434+0000 7f51859bc640 -1 osd.2 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.434 INFO:tasks.ceph.osd.0.vm02.stderr:2026-04-17T12:41:44.434+0000 7fe101328640 -1 osd.0 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.434 INFO:tasks.ceph.osd.3.vm02.stderr:2026-04-17T12:41:44.434+0000 7f217f188640 -1 osd.3 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.435 INFO:tasks.ceph.osd.1.vm02.stderr:2026-04-17T12:41:44.434+0000 7f583ef9c640 -1 osd.1 71 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-04-17T12:41:44.629 INFO:tasks.ceph.mgr.y.vm02.stderr:daemon-helper: command crashed with signal 15 2026-04-17T12:42:14.419 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_large_scale 2026-04-17T12:42:14.419 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------- live log call --------------------------------- 2026-04-17T12:42:14.419 INFO:teuthology.orchestra.run.vm02.stdout:WARNING dedup.test_dedup:test_dedup.py:2748 test_dedup_dry_large_scale: failed!! 2026-04-17T12:42:19.803 INFO:teuthology.orchestra.run.vm02.stdout:FAILED [ 97%] 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_cleanup PASSED [100%] 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:=================================== FAILURES =================================== 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:__________________________ test_dedup_dry_large_scale __________________________ 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:self = 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: def _new_conn(self): 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: """Establish a socket connection and set nodelay settings on it. 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: :return: New socket connection. 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: """ 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: extra_kw = {} 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: if self.source_address: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: extra_kw["source_address"] = self.source_address 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: if self.socket_options: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: extra_kw["socket_options"] = self.socket_options 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:> conn = connection.create_connection( 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: (self._dns_host, self.port), self.timeout, **extra_kw 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:174: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py:95: in create_connection 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: raise err 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:address = ('vm02.local', 80), timeout = 60, source_address = None 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout:socket_options = [(6, 1, 1)] 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: def create_connection( 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: address, 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: timeout=socket._GLOBAL_DEFAULT_TIMEOUT, 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: source_address=None, 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: socket_options=None, 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: ): 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: """Connect to *address* and return the socket object. 2026-04-17T12:42:19.806 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: Convenience function. Connect to *address* (a 2-tuple ``(host, 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: port)``) and return the socket object. Passing the optional 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: *timeout* parameter will set the timeout on the socket instance 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: before attempting to connect. If no *timeout* is supplied, the 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: global default timeout setting returned by :func:`socket.getdefaulttimeout` 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: is used. If *source_address* is set it must be a tuple of (host, port) 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: for the socket to bind as a source address before making the connection. 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: An host of '' or port 0 tells the OS to use the default. 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: """ 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: host, port = address 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: if host.startswith("["): 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: host = host.strip("[]") 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: err = None 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: # Using the value from allowed_gai_family() in the context of getaddrinfo lets 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: # us select whether to work with IPv4 DNS records, IPv6 records, or both. 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: # The original create_connection function always returns all records. 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: family = allowed_gai_family() 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: host.encode("idna") 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: except UnicodeError: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: return six.raise_from( 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: LocationParseError(u"'%s', label empty or too long" % host), None 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM): 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: af, socktype, proto, canonname, sa = res 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: sock = None 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: sock = socket.socket(af, socktype, proto) 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: # If provided, set socket level options before connecting. 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: _set_socket_options(sock, socket_options) 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: sock.settimeout(timeout) 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: if source_address: 2026-04-17T12:42:19.807 INFO:teuthology.orchestra.run.vm02.stdout: sock.bind(source_address) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:> sock.connect(sa) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:E ConnectionRefusedError: [Errno 111] Connection refused 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py:85: ConnectionRefusedError 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:During handling of the above exception, another exception occurred: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:self = 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:request = 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: def send(self, request): 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: proxy_url = self._proxy_config.proxy_url_for(request.url) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: manager = self._get_connection_manager(request.url, proxy_url) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: conn = manager.connection_from_url(request.url) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: self._setup_ssl_cert(conn, request.url, self._verify) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: if ensure_boolean( 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: os.environ.get('BOTO_EXPERIMENTAL__ADD_PROXY_HOST_HEADER', '') 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: ): 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: # This is currently an "experimental" feature which provides 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: # no guarantees of backwards compatibility. It may be subject 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: # to change or removal in any patch version. Anyone opting in 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: # to this feature should strictly pin botocore. 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: host = urlparse(request.url).hostname 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: conn.proxy_headers['host'] = host 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: request_target = self._get_request_target(request.url, proxy_url) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:> urllib_response = conn.urlopen( 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: method=request.method, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: url=request_target, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: body=request.body, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: headers=request.headers, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: retries=Retry(False), 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: assert_same_host=False, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: preload_content=False, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: decode_content=False, 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: chunked=self._chunked(request.headers), 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/httpsession.py:477: 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py:802: in urlopen 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: retries = retries.increment( 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/util/retry.py:527: in increment 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: raise six.reraise(type(error), error, _stacktrace) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/packages/six.py:770: in reraise 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: raise value 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py:716: in urlopen 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: httplib_response = self._make_request( 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py:416: in _make_request 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: conn.request(method, url, **httplib_request_kw) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py:96: in request 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: rval = super().request(method, url, body, headers, *args, **kwargs) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:244: in request 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: super(HTTPConnection, self).request(method, url, body=body, headers=headers) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:/usr/lib64/python3.9/http/client.py:1285: in request 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: self._send_request(method, url, body, headers, encode_chunked) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:/usr/lib64/python3.9/http/client.py:1331: in _send_request 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout: self.endheaders(body, encode_chunked=encode_chunked) 2026-04-17T12:42:19.808 INFO:teuthology.orchestra.run.vm02.stdout:/usr/lib64/python3.9/http/client.py:1280: in endheaders 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: self._send_output(message_body, encode_chunked=encode_chunked) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py:123: in _send_output 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: self.send(msg) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py:223: in send 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: return super().send(str) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:/usr/lib64/python3.9/http/client.py:980: in send 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: self.connect() 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:205: in connect 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: conn = self._new_conn() 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:self = 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: def _new_conn(self): 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: """Establish a socket connection and set nodelay settings on it. 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: :return: New socket connection. 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: """ 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: extra_kw = {} 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: if self.source_address: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: extra_kw["source_address"] = self.source_address 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: if self.socket_options: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: extra_kw["socket_options"] = self.socket_options 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: conn = connection.create_connection( 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: (self._dns_host, self.port), self.timeout, **extra_kw 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: except SocketTimeout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: raise ConnectTimeoutError( 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: self, 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: "Connection to %s timed out. (connect timeout=%s)" 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: % (self.host, self.timeout), 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: except SocketError as e: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:> raise NewConnectionError( 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: self, "Failed to establish a new connection: %s" % e 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:E urllib3.exceptions.NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/urllib3/connection.py:186: NewConnectionError 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout:During handling of the above exception, another exception occurred: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: @pytest.mark.basic_test 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: def test_dedup_dry_large_scale(): 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: #return 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: prepare_test() 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: max_copies_count=3 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: num_threads=64 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: num_files=32*1024 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: size=1*KB 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: files=[] 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: config=TransferConfig(multipart_threshold=size, multipart_chunksize=1*MB) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: log.debug("test_dedup_dry_large_scale_new: connect to AWS ...") 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: gen_files_fixed_size(files, num_files, size, max_copies_count) 2026-04-17T12:42:19.809 INFO:teuthology.orchestra.run.vm02.stdout: conns=get_connections(num_threads) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: bucket_names=get_buckets(num_threads) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: for i in range(num_threads): 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: conns[i].create_bucket(Bucket=bucket_names[i]) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: threads_simple_dedup_with_tenants(files, conns, bucket_names, config, True) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: except: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: log.warning("test_dedup_dry_large_scale: failed!!") 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: finally: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: # cleanup must be executed even after a failure 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:> cleanup_all_buckets(bucket_names, conns) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py:2751: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py:496: in cleanup_all_buckets 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: delete_bucket_with_all_objects(bucket_name, conn) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py:452: in delete_bucket_with_all_objects 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: listing=conn.list_objects(Bucket=bucket_name, Marker=marker, MaxKeys=max_keys) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/client.py:606: in _api_call 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._make_api_call(operation_name, kwargs) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/context.py:123: in wrapper 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return func(*args, **kwargs) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/client.py:1076: in _make_api_call 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: http, parsed_response = self._make_request( 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/client.py:1100: in _make_request 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._endpoint.make_request(operation_model, request_dict) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:119: in make_request 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._send_request(request_dict, operation_model) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:200: in _send_request 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: while self._needs_retry( 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:360: in _needs_retry 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: responses = self._event_emitter.emit( 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/hooks.py:412: in emit 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._emitter.emit(aliased_event_name, **kwargs) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/hooks.py:256: in emit 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._emit(event_name, kwargs) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/hooks.py:239: in _emit 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: response = handler(**kwargs) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:207: in __call__ 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: if self._checker(**checker_kwargs): 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:284: in __call__ 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: should_retry = self._should_retry( 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:320: in _should_retry 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._checker(attempt_number, response, caught_exception) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:363: in __call__ 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: checker_response = checker( 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:247: in __call__ 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self._check_caught_exception( 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/retryhandler.py:416: in _check_caught_exception 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: raise caught_exception 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:279: in _do_get_response 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: http_response = self._send(request) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/endpoint.py:383: in _send 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: return self.http_session.send(request) 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:self = 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout:request = 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: def send(self, request): 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: try: 2026-04-17T12:42:19.810 INFO:teuthology.orchestra.run.vm02.stdout: proxy_url = self._proxy_config.proxy_url_for(request.url) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: manager = self._get_connection_manager(request.url, proxy_url) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: conn = manager.connection_from_url(request.url) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: self._setup_ssl_cert(conn, request.url, self._verify) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: if ensure_boolean( 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: os.environ.get('BOTO_EXPERIMENTAL__ADD_PROXY_HOST_HEADER', '') 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: ): 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # This is currently an "experimental" feature which provides 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # no guarantees of backwards compatibility. It may be subject 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # to change or removal in any patch version. Anyone opting in 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # to this feature should strictly pin botocore. 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: host = urlparse(request.url).hostname 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: conn.proxy_headers['host'] = host 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: request_target = self._get_request_target(request.url, proxy_url) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: urllib_response = conn.urlopen( 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: method=request.method, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: url=request_target, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: body=request.body, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: headers=request.headers, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: retries=Retry(False), 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: assert_same_host=False, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: preload_content=False, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: decode_content=False, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: chunked=self._chunked(request.headers), 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: http_response = botocore.awsrequest.AWSResponse( 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: request.url, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: urllib_response.status, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: urllib_response.headers, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: urllib_response, 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: ) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: if not request.stream_output: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # Cause the raw stream to be exhausted immediately. We do it 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # this way instead of using preload_content because 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: # preload_content will never buffer chunked responses 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: http_response.content 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: return http_response 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: except URLLib3SSLError as e: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: raise SSLError(endpoint_url=request.url, error=e) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: except (NewConnectionError, socket.gaierror) as e: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout:> raise EndpointConnectionError(endpoint_url=request.url, error=e) 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout:E botocore.exceptions.EndpointConnectionError: Could not connect to the endpoint URL: "http://vm02.local:80/bfgmloszdkgaowak-86?marker=&max-keys=1000&encoding-type=url" 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout:.tox/py/lib/python3.9/site-packages/botocore/httpsession.py:506: EndpointConnectionError 2026-04-17T12:42:19.811 INFO:teuthology.orchestra.run.vm02.stdout:----------------------------- Captured stderr call ----------------------------- 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:ignoring --setuser ceph since I am not root 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:ignoring --setgroup ceph since I am not root 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:ignoring --setuser ceph since I am not root 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:ignoring --setgroup ceph since I am not root 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:Process Process-87: 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:Traceback (most recent call last): 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 174, in _new_conn 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: conn = connection.create_connection( 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py", line 95, in create_connection 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: raise err 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/util/connection.py", line 85, in create_connection 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: sock.connect(sa) 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:ConnectionRefusedError: [Errno 111] Connection refused 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:During handling of the above exception, another exception occurred: 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout:Traceback (most recent call last): 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/httpsession.py", line 477, in send 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: urllib_response = conn.urlopen( 2026-04-17T12:42:20.071 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py", line 802, in urlopen 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: retries = retries.increment( 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/util/retry.py", line 527, in increment 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: raise six.reraise(type(error), error, _stacktrace) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/packages/six.py", line 770, in reraise 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: raise value 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py", line 716, in urlopen 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: httplib_response = self._make_request( 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connectionpool.py", line 416, in _make_request 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: conn.request(method, url, **httplib_request_kw) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py", line 96, in request 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: rval = super().request(method, url, body, headers, *args, **kwargs) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 244, in request 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: super(HTTPConnection, self).request(method, url, body=body, headers=headers) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/usr/lib64/python3.9/http/client.py", line 1285, in request 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self._send_request(method, url, body, headers, encode_chunked) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/usr/lib64/python3.9/http/client.py", line 1331, in _send_request 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self.endheaders(body, encode_chunked=encode_chunked) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/usr/lib64/python3.9/http/client.py", line 1280, in endheaders 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self._send_output(message_body, encode_chunked=encode_chunked) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py", line 123, in _send_output 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self.send(msg) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/botocore/awsrequest.py", line 223, in send 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: return super().send(str) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/usr/lib64/python3.9/http/client.py", line 980, in send 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self.connect() 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 205, in connect 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: conn = self._new_conn() 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/urllib3/connection.py", line 186, in _new_conn 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: raise NewConnectionError( 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:urllib3.exceptions.NewConnectionError: : Failed to establish a new connection: [Errno 111] Connection refused 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:During handling of the above exception, another exception occurred: 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:Traceback (most recent call last): 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/usr/lib64/python3.9/multiprocessing/process.py", line 315, in _bootstrap 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self.run() 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/usr/lib64/python3.9/multiprocessing/process.py", line 108, in run 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: self._target(*self._args, **self._kwargs) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: File "/home/ubuntu/cephtest/ceph/src/test/rgw/dedup/test_dedup.py 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:------------------------------ Captured log call ------------------------------- 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:WARNING dedup.test_dedup:test_dedup.py:2748 test_dedup_dry_large_scale: failed!! 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:=============================== warnings summary =============================== 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_with_tenants 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_multipart 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_large_mix 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_basic_with_tenants 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_multipart_with_tenants 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_small_multipart_with_tenants 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_large_scale_with_tenants 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:test_dedup.py::test_dedup_dry_large_scale 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/.tox/py/lib/python3.9/site-packages/boto3/compat.py:89: PythonDeprecationWarning: Boto3 will no longer support Python 3.9 starting April 29, 2026. To continue receiving service updates, bug fixes, and security updates please upgrade to Python 3.10 or later. More information can be found here: https://aws.amazon.com/blogs/developer/python-support-policy-updates-for-aws-sdks-and-tools/ 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: warnings.warn(warning, PythonDeprecationWarning) 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:-- Docs: https://docs.pytest.org/en/stable/how-to/capture-warnings.html 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:=========================== short test summary info ============================ 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:FAILED test_dedup.py::test_dedup_dry_large_scale - botocore.exceptions.Endpoi... 2026-04-17T12:42:20.072 INFO:teuthology.orchestra.run.vm02.stdout:============ 1 failed, 33 passed, 8 warnings in 1292.41s (0:21:32) ============= 2026-04-17T12:42:20.262 INFO:teuthology.orchestra.run.vm02.stdout:py: exit 1 (1292.95 seconds) /home/ubuntu/cephtest/ceph/src/test/rgw/dedup> pytest -v -m 'basic_test or request_test or example_test' pid=67521 2026-04-17T12:42:20.263 INFO:teuthology.orchestra.run.vm02.stdout: py: FAIL code 1 (1296.11=setup[3.16]+cmd[1292.95] seconds) 2026-04-17T12:42:20.263 INFO:teuthology.orchestra.run.vm02.stdout: evaluation failed :( (1296.13 seconds) 2026-04-17T12:42:20.283 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:42:20.283 ERROR:teuthology.contextutil:Saw exception from nested tasks Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/contextutil.py", line 30, in nested vars.append(enter()) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 191, in run_tests toxvenv_sh(ctx, remote, args, label="dedup tests against rgw") File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 165, in toxvenv_sh return remote.sh(['source', activate, run.Raw('&&')] + args, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 97, in sh proc = self.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed (dedup tests against rgw) on vm02 with status 1: "source /home/ubuntu/cephtest/tox-venv/bin/activate && cd /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/ && DEDUPTESTS_CONF=./deduptests.client.0.conf tox -- -v -m 'basic_test or request_test or example_test'" 2026-04-17T12:42:20.284 INFO:tasks.dedup_tests:Removing dedup-tests.conf file... 2026-04-17T12:42:20.284 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/deduptests.client.0.conf 2026-04-17T12:42:20.312 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user rm --uid foo.client.0 --purge-data --cluster ceph 2026-04-17T12:42:20.417 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:42:20.417 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:47:20.420 INFO:teuthology.orchestra.run.vm02.stderr:failed to fetch mon config (--no-mon-config to skip) 2026-04-17T12:47:20.423 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:47:20.423 INFO:tasks.dedup_tests:Removing dedup-tests... 2026-04-17T12:47:20.423 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/ceph 2026-04-17T12:47:20.981 ERROR:teuthology.run_tasks:Saw exception from tasks. Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 107, in create_users yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 30, in nested vars.append(enter()) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 191, in run_tests toxvenv_sh(ctx, remote, args, label="dedup tests against rgw") File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 165, in toxvenv_sh return remote.sh(['source', activate, run.Raw('&&')] + args, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 97, in sh proc = self.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed (dedup tests against rgw) on vm02 with status 1: "source /home/ubuntu/cephtest/tox-venv/bin/activate && cd /home/ubuntu/cephtest/ceph/src/test/rgw/dedup/ && DEDUPTESTS_CONF=./deduptests.client.0.conf tox -- -v -m 'basic_test or request_test or example_test'" During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 112, in run_tasks manager.__enter__() File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 240, in task with contextutil.nested( File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 135, in __enter__ return next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 45, in download yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/dedup_tests.py", line 114, in create_users ctx.cluster.only(client).run( File "/home/teuthos/teuthology/teuthology/orchestra/cluster.py", line 85, in run procs = [remote.run(**kwargs, wait=_wait) for remote in remotes] File "/home/teuthos/teuthology/teuthology/orchestra/cluster.py", line 85, in procs = [remote.run(**kwargs, wait=_wait) for remote in remotes] File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user rm --uid foo.client.0 --purge-data --cluster ceph' 2026-04-17T12:47:20.981 DEBUG:teuthology.run_tasks:Unwinding manager dedup-tests 2026-04-17T12:47:20.983 DEBUG:teuthology.run_tasks:Unwinding manager tox 2026-04-17T12:47:20.985 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/tox-venv 2026-04-17T12:47:21.052 DEBUG:teuthology.run_tasks:Unwinding manager tox 2026-04-17T12:47:21.054 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/tox-venv 2026-04-17T12:47:21.069 DEBUG:teuthology.run_tasks:Unwinding manager rgw 2026-04-17T12:47:21.071 DEBUG:tasks.rgw.client.0:waiting for process to exit 2026-04-17T12:47:21.071 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:47:21.071 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:47:21.071 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: "sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper term radosgw --rgw-frontends 'beast port=80' -n client.0 --cluster ceph -k /etc/ceph/ceph.client.0.keyring --log-file /var/log/ceph/rgw.ceph.client.0.log --rgw_ops_log_socket_path /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock --foreground | sudo tee /var/log/ceph/rgw.ceph.client.0.stdout 2>&1" 2026-04-17T12:47:21.072 INFO:tasks.rgw.client.0:Stopped 2026-04-17T12:47:21.072 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/rgw.opslog.ceph.client.0.sock 2026-04-17T12:47:21.127 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/vault-root-token 2026-04-17T12:47:21.199 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /home/ubuntu/cephtest/url_file 2026-04-17T12:47:21.264 INFO:tasks.util.rgw:rgwadmin: client.0 : ['gc', 'process', '--include-all'] 2026-04-17T12:47:21.264 DEBUG:tasks.util.rgw:rgwadmin: cmd=['adjust-ulimits', 'ceph-coverage', '/home/ubuntu/cephtest/archive/coverage', 'radosgw-admin', '--log-to-stderr', '--format', 'json', '-n', 'client.0', '--cluster', 'ceph', 'gc', 'process', '--include-all'] 2026-04-17T12:47:21.264 DEBUG:teuthology.orchestra.run.vm02:> adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all 2026-04-17T12:47:21.338 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setuser ceph since I am not root 2026-04-17T12:47:21.338 INFO:teuthology.orchestra.run.vm02.stderr:ignoring --setgroup ceph since I am not root 2026-04-17T12:52:21.340 INFO:teuthology.orchestra.run.vm02.stderr:2026-04-17T12:52:21.340+0000 7f655e824900 0 monclient(hunting): authenticate timed out after 300 2026-04-17T12:52:21.340 INFO:teuthology.orchestra.run.vm02.stderr:failed to fetch mon config (--no-mon-config to skip) 2026-04-17T12:52:21.342 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:52:21.342 ERROR:teuthology.run_tasks:Manager failed: rgw Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' 2026-04-17T12:52:21.342 DEBUG:teuthology.run_tasks:Unwinding manager openssl_keys 2026-04-17T12:52:21.344 DEBUG:teuthology.run_tasks:Unwinding manager ceph 2026-04-17T12:52:21.346 INFO:tasks.ceph.ceph_manager.ceph:waiting for clean 2026-04-17T12:52:21.347 DEBUG:teuthology.orchestra.run.vm02:> sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json 2026-04-17T12:54:21.412 DEBUG:teuthology.orchestra.run:got remote process result: 124 2026-04-17T12:54:21.413 ERROR:teuthology.contextutil:Saw exception from nested tasks Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' 2026-04-17T12:54:21.413 INFO:teuthology.misc:Shutting down mds daemons... 2026-04-17T12:54:21.413 INFO:teuthology.misc:Shutting down osd daemons... 2026-04-17T12:54:21.413 DEBUG:tasks.ceph.osd.0:waiting for process to exit 2026-04-17T12:54:21.413 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.0:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.1:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.1:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.2:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.2:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.3:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.3:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.4:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.4:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.5:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.5:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.6:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.6:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.osd.7:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.osd.7:Stopped 2026-04-17T12:54:21.414 INFO:teuthology.misc:Shutting down mgr daemons... 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.mgr.y:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:54:21.414 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mgr -f --cluster ceph -i y' 2026-04-17T12:54:21.414 INFO:tasks.ceph.mgr.y:Stopped 2026-04-17T12:54:21.414 DEBUG:tasks.ceph.mgr.x:waiting for process to exit 2026-04-17T12:54:21.414 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.414 INFO:tasks.ceph.mgr.x:Stopped 2026-04-17T12:54:21.414 INFO:teuthology.misc:Shutting down mon daemons... 2026-04-17T12:54:21.415 DEBUG:tasks.ceph.mon.a:waiting for process to exit 2026-04-17T12:54:21.415 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.415 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:54:21.415 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i a' 2026-04-17T12:54:21.415 INFO:tasks.ceph.mon.a:Stopped 2026-04-17T12:54:21.415 DEBUG:tasks.ceph.mon.c:waiting for process to exit 2026-04-17T12:54:21.415 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.415 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:54:21.415 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i c' 2026-04-17T12:54:21.415 INFO:tasks.ceph.mon.c:Stopped 2026-04-17T12:54:21.415 DEBUG:tasks.ceph.mon.b:waiting for process to exit 2026-04-17T12:54:21.415 INFO:teuthology.orchestra.run:waiting for 300 2026-04-17T12:54:21.415 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:54:21.415 ERROR:teuthology.orchestra.daemon.state:Error while waiting for process to exit Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 1526, in run_daemon yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/orchestra/daemon/state.py", line 146, in stop run.wait([self.proc], timeout=timeout) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage daemon-helper kill ceph-mon -f --cluster ceph -i b' 2026-04-17T12:54:21.415 INFO:tasks.ceph.mon.b:Stopped 2026-04-17T12:54:21.415 INFO:tasks.ceph:Checking cluster log for badness... 2026-04-17T12:54:21.415 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v '\(PG_AVAILABILITY\)' | egrep -v '\(PG_DEGRADED\)' | egrep -v '\(POOL_APP_NOT_ENABLED\)' | egrep -v 'not have an application enabled' | head -n 1 2026-04-17T12:54:21.442 INFO:teuthology.orchestra.run.vm02.stdout:2026-04-17T12:30:22.726105+0000 mon.a (mon.0) 664 : cluster [ERR] Health check failed: mon c is very low on available space (MON_DISK_CRIT) 2026-04-17T12:54:21.443 WARNING:tasks.ceph:Found errors (ERR|WRN|SEC) in cluster log 2026-04-17T12:54:21.443 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-0 on ubuntu@vm02.local 2026-04-17T12:54:21.443 DEBUG:teuthology.orchestra.run.vm02:> sync && sudo umount -f /var/lib/ceph/osd/ceph-0 2026-04-17T12:54:21.564 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-1 on ubuntu@vm02.local 2026-04-17T12:54:21.564 DEBUG:teuthology.orchestra.run.vm02:> sync && sudo umount -f /var/lib/ceph/osd/ceph-1 2026-04-17T12:54:21.647 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-2 on ubuntu@vm02.local 2026-04-17T12:54:21.647 DEBUG:teuthology.orchestra.run.vm02:> sync && sudo umount -f /var/lib/ceph/osd/ceph-2 2026-04-17T12:54:21.726 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-3 on ubuntu@vm02.local 2026-04-17T12:54:21.726 DEBUG:teuthology.orchestra.run.vm02:> sync && sudo umount -f /var/lib/ceph/osd/ceph-3 2026-04-17T12:54:21.807 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-4 on ubuntu@vm06.local 2026-04-17T12:54:21.807 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-4 2026-04-17T12:54:21.924 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-5 on ubuntu@vm06.local 2026-04-17T12:54:21.924 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-5 2026-04-17T12:54:22.022 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-6 on ubuntu@vm06.local 2026-04-17T12:54:22.022 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-6 2026-04-17T12:54:22.130 INFO:tasks.ceph:Unmounting /var/lib/ceph/osd/ceph-7 on ubuntu@vm06.local 2026-04-17T12:54:22.130 DEBUG:teuthology.orchestra.run.vm06:> sync && sudo umount -f /var/lib/ceph/osd/ceph-7 2026-04-17T12:54:22.222 INFO:tasks.ceph:Archiving mon data... 2026-04-17T12:54:22.222 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/mon/ceph-a to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/data/mon.a.tgz 2026-04-17T12:54:22.222 DEBUG:teuthology.orchestra.run.vm02:> mktemp 2026-04-17T12:54:22.238 INFO:teuthology.orchestra.run.vm02.stdout:/tmp/tmp.1z7JrMQGZ9 2026-04-17T12:54:22.238 DEBUG:teuthology.orchestra.run.vm02:> sudo tar cz -f - -C /var/lib/ceph/mon/ceph-a -- . > /tmp/tmp.1z7JrMQGZ9 2026-04-17T12:54:22.374 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0666 /tmp/tmp.1z7JrMQGZ9 2026-04-17T12:54:22.457 DEBUG:teuthology.orchestra.remote:vm02:/tmp/tmp.1z7JrMQGZ9 is 509KB 2026-04-17T12:54:22.517 DEBUG:teuthology.orchestra.run.vm02:> rm -fr /tmp/tmp.1z7JrMQGZ9 2026-04-17T12:54:22.532 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/mon/ceph-c to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/data/mon.c.tgz 2026-04-17T12:54:22.532 DEBUG:teuthology.orchestra.run.vm02:> mktemp 2026-04-17T12:54:22.589 INFO:teuthology.orchestra.run.vm02.stdout:/tmp/tmp.3JGHlz1NdV 2026-04-17T12:54:22.589 DEBUG:teuthology.orchestra.run.vm02:> sudo tar cz -f - -C /var/lib/ceph/mon/ceph-c -- . > /tmp/tmp.3JGHlz1NdV 2026-04-17T12:54:22.731 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0666 /tmp/tmp.3JGHlz1NdV 2026-04-17T12:54:22.809 DEBUG:teuthology.orchestra.remote:vm02:/tmp/tmp.3JGHlz1NdV is 532KB 2026-04-17T12:54:22.868 DEBUG:teuthology.orchestra.run.vm02:> rm -fr /tmp/tmp.3JGHlz1NdV 2026-04-17T12:54:22.882 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/mon/ceph-b to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/data/mon.b.tgz 2026-04-17T12:54:22.882 DEBUG:teuthology.orchestra.run.vm06:> mktemp 2026-04-17T12:54:22.898 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:54:22.898 INFO:teuthology.orchestra.run.vm06.stderr:mktemp: failed to create file via template ‘/tmp/tmp.XXXXXXXXXX’: No space left on device 2026-04-17T12:54:22.938 INFO:teuthology.util.scanner:summary_data or yaml_file is empty! 2026-04-17T12:54:22.953 INFO:teuthology.util.scanner:summary_data or yaml_file is empty! 2026-04-17T12:54:22.971 INFO:teuthology.util.scanner:summary_data or yaml_file is empty! 2026-04-17T12:54:22.971 INFO:tasks.ceph:Archiving crash dumps... 2026-04-17T12:54:22.971 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/crash to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/remote/vm02/crash 2026-04-17T12:54:22.971 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/crash -- . 2026-04-17T12:54:23.000 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/crash to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/remote/vm06/crash 2026-04-17T12:54:23.001 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/crash -- . 2026-04-17T12:54:23.023 DEBUG:teuthology.misc:Transferring archived files from vm08:/var/lib/ceph/crash to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/remote/vm08/crash 2026-04-17T12:54:23.023 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /var/lib/ceph/crash -- . 2026-04-17T12:54:23.054 INFO:tasks.ceph:Compressing logs... 2026-04-17T12:54:23.054 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-04-17T12:54:23.056 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-04-17T12:54:23.066 DEBUG:teuthology.orchestra.run.vm08:> time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-04-17T12:54:23.078 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.tmp-client.admin.53724.log 2026-04-17T12:54:23.079 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.0.log 2026-04-17T12:54:23.079 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph.tmp-client.admin.53724.log: 0.0% -- replaced with /var/log/ceph/ceph.tmp-client.admin.53724.log.gz 2026-04-17T12:54:23.079 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.1.log 2026-04-17T12:54:23.079 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-osd.0.log: gzip -5 --verbose -- /var/log/ceph/ceph-osd.2.log 2026-04-17T12:54:23.080 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5/var/log/ceph/ceph-osd.1.log: --verbose -- /var/log/ceph/ceph-osd.3.log 2026-04-17T12:54:23.085 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-osd.2.log: /var/log/ceph/ceph-osd.3.log: gzip -5 --verbose -- /var/log/ceph/ceph-mon.a.log 2026-04-17T12:54:23.085 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mon.c.log 2026-04-17T12:54:23.087 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.4.log 2026-04-17T12:54:23.088 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.5.log 2026-04-17T12:54:23.088 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-osd.6.log 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-osd.4.log.gz: No space left on device 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr:gzipgzip: /var/log/ceph/ceph-osd.5.log.gz: No space left on device 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr: -5 --verbose -- /var/log/ceph/ceph-osd.7.log 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-osd.6.log.gz: No space left on device 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mon.b.log 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-osd.7.log.gz: No space left on device 2026-04-17T12:54:23.089 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.log 2026-04-17T12:54:23.090 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-mon.b.log.gz: No space left on device 2026-04-17T12:54:23.090 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mgr.x.log 2026-04-17T12:54:23.090 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph.log.gz: No space left on device 2026-04-17T12:54:23.090 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59337.log 2026-04-17T12:54:23.090 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-mgr.x.log.gz: No space left on device 2026-04-17T12:54:23.090 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59385.log 2026-04-17T12:54:23.091 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59337.log.gz: No space left on device 2026-04-17T12:54:23.091 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph.audit.log 2026-04-17T12:54:23.091 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59385.log.gz: No space left on device 2026-04-17T12:54:23.091 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59433.log 2026-04-17T12:54:23.091 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph.audit.log.gz: No space left on device 2026-04-17T12:54:23.091 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59481.log 2026-04-17T12:54:23.092 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59433.log.gz: No space left on device 2026-04-17T12:54:23.092 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59529.log 2026-04-17T12:54:23.092 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59577.log 2026-04-17T12:54:23.092 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59529.log.gz: No space left on device 2026-04-17T12:54:23.092 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59481.log.gz: No space left on device 2026-04-17T12:54:23.092 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59625.log 2026-04-17T12:54:23.093 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59577.log.gz: No space left on device 2026-04-17T12:54:23.093 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.59673.log 2026-04-17T12:54:23.093 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59625.log.gz: No space left on device 2026-04-17T12:54:23.093 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63329.log 2026-04-17T12:54:23.093 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.59673.log.gz: No space left on device 2026-04-17T12:54:23.093 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63377.log 2026-04-17T12:54:23.094 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.63329.log.gz: No space left on device 2026-04-17T12:54:23.094 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63425.log 2026-04-17T12:54:23.094 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.63377.log.gz: No space left on device 2026-04-17T12:54:23.094 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63473.log 2026-04-17T12:54:23.094 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.63425.log.gz: No space left on device 2026-04-17T12:54:23.094 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.63496.log 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.admin.63473.log.gz: No space left on device 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.63604.log 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.1.63496.log.gz: No space left on device 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.63707.log 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.1.63604.log.gz: No space left on device 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.63810.log 2026-04-17T12:54:23.095 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.1.63707.log.gz: No space left on device 2026-04-17T12:54:23.096 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.1.63913.log 2026-04-17T12:54:23.096 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.1.63810.log.gz: No space left on device 2026-04-17T12:54:23.096 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/rgw.ceph.client.1.log 2026-04-17T12:54:23.096 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ceph-client.1.63913.log.gz: No space left on device 2026-04-17T12:54:23.096 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ops-log-ceph-client.1.log 2026-04-17T12:54:23.096 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/rgw.ceph.client.1.log.gz: No space left on device 2026-04-17T12:54:23.097 INFO:teuthology.orchestra.run.vm06.stderr:gzip: /var/log/ceph/ops-log-ceph-client.1.log.gz: No space left on device 2026-04-17T12:54:23.098 INFO:teuthology.orchestra.run.vm06.stderr: 2026-04-17T12:54:23.098 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.019s 2026-04-17T12:54:23.098 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.014s 2026-04-17T12:54:23.098 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.018s 2026-04-17T12:54:23.100 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-mon.a.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.60165.log 2026-04-17T12:54:23.122 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-mon.c.log: gzip -5 --verbose -- /var/log/ceph/ceph.log 2026-04-17T12:54:23.122 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.52458.log 2026-04-17T12:54:23.122 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.60165.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.60165.log.gz 2026-04-17T12:54:23.123 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.52506.log 2026-04-17T12:54:23.123 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.52554.log 2026-04-17T12:54:23.123 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.admin.52458.log: /var/log/ceph/ceph-client.admin.52506.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.52458.log.gz 2026-04-17T12:54:23.124 INFO:teuthology.orchestra.run.vm08.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.52506.log.gz 2026-04-17T12:54:23.124 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.52602.log 2026-04-17T12:54:23.124 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.admin.52554.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.2.52627.log 2026-04-17T12:54:23.124 INFO:teuthology.orchestra.run.vm08.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.52554.log.gz 2026-04-17T12:54:23.124 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.admin.52602.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.2.52735.log 2026-04-17T12:54:23.124 INFO:teuthology.orchestra.run.vm08.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.52602.log.gz 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.2.52838.log 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.2.52627.log: /var/log/ceph/ceph-client.2.52735.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.2.52941.log 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr: 45.3% -- replaced with /var/log/ceph/ceph-client.2.52735.log.gz 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr: 83.2% -- replaced with /var/log/ceph/ceph-client.2.52627.log.gz 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.2.53044.log 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.2.52838.log: 44.9% -- replaced with /var/log/ceph/ceph-client.2.52838.log.gz 2026-04-17T12:54:23.125 INFO:teuthology.orchestra.run.vm08.stderr:gzip -5 --verbose -- /var/log/ceph/rgw.ceph.client.2.log 2026-04-17T12:54:23.126 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.2.52941.log: 45.6% -- replaced with /var/log/ceph/ceph-client.2.52941.log.gz 2026-04-17T12:54:23.126 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/ceph-client.2.53044.log: gzip -5 --verbose -- /var/log/ceph/ops-log-ceph-client.2.log 2026-04-17T12:54:23.126 INFO:teuthology.orchestra.run.vm08.stderr: 45.6% -- replaced with /var/log/ceph/ceph-client.2.53044.log.gz 2026-04-17T12:54:23.126 INFO:teuthology.orchestra.run.vm08.stderr:/var/log/ceph/rgw.ceph.client.2.log: /var/log/ceph/ops-log-ceph-client.2.log: 35.1% -- replaced with /var/log/ceph/ops-log-ceph-client.2.log.gz 2026-04-17T12:54:23.135 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-mgr.y.log 2026-04-17T12:54:23.137 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph.log: 92.8% -- replaced with /var/log/ceph/ceph.log.gz 2026-04-17T12:54:23.141 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.60234.log 2026-04-17T12:54:23.147 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-mgr.y.log: gzip -5 --verbose -- /var/log/ceph/ceph.audit.log 2026-04-17T12:54:23.147 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.60234.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.60234.log.gz 2026-04-17T12:54:23.153 INFO:teuthology.orchestra.run.vm02.stderr: 94.5% -- replaced with /var/log/ceph/ceph-mgr.y.log.gz 2026-04-17T12:54:23.154 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.60508.log 2026-04-17T12:54:23.157 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph.audit.log: 94.4% -- replaced with /var/log/ceph/ceph.audit.log.gz 2026-04-17T12:54:23.160 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63734.log 2026-04-17T12:54:23.160 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.60508.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.60508.log.gz 2026-04-17T12:54:23.166 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63829.log 2026-04-17T12:54:23.166 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.63734.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.63734.log.gz 2026-04-17T12:54:23.178 INFO:teuthology.orchestra.run.vm08.stderr: 91.2% -- replaced with /var/log/ceph/rgw.ceph.client.2.log.gz 2026-04-17T12:54:23.180 INFO:teuthology.orchestra.run.vm08.stderr: 2026-04-17T12:54:23.180 INFO:teuthology.orchestra.run.vm08.stderr:real 0m0.067s 2026-04-17T12:54:23.180 INFO:teuthology.orchestra.run.vm08.stderr:user 0m0.061s 2026-04-17T12:54:23.180 INFO:teuthology.orchestra.run.vm08.stderr:sys 0m0.017s 2026-04-17T12:54:23.181 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63903.log 2026-04-17T12:54:23.181 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.63829.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.63829.log.gz 2026-04-17T12:54:23.187 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.63930.log 2026-04-17T12:54:23.187 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.63903.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.63903.log.gz 2026-04-17T12:54:23.196 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64021.log 2026-04-17T12:54:23.196 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.63930.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.63930.log.gz 2026-04-17T12:54:23.201 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64071.log 2026-04-17T12:54:23.201 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64021.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64021.log.gz 2026-04-17T12:54:23.212 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64122.log 2026-04-17T12:54:23.212 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64071.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64071.log.gz 2026-04-17T12:54:23.217 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64172.log 2026-04-17T12:54:23.217 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64122.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64122.log.gz 2026-04-17T12:54:23.228 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64410.log 2026-04-17T12:54:23.228 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64172.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64172.log.gz 2026-04-17T12:54:23.234 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64417.log 2026-04-17T12:54:23.234 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64410.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64410.log.gz 2026-04-17T12:54:23.243 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64408.log 2026-04-17T12:54:23.243 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64417.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64417.log.gz 2026-04-17T12:54:23.249 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64398.log 2026-04-17T12:54:23.249 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64408.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64408.log.gz 2026-04-17T12:54:23.263 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64402.log 2026-04-17T12:54:23.263 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64398.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64398.log.gz 2026-04-17T12:54:23.268 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64413.log 2026-04-17T12:54:23.270 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64402.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64402.log.gz 2026-04-17T12:54:23.277 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64418.log 2026-04-17T12:54:23.277 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64413.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64413.log.gz 2026-04-17T12:54:23.282 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64415.log 2026-04-17T12:54:23.282 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64418.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64418.log.gz 2026-04-17T12:54:23.287 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64694.log 2026-04-17T12:54:23.287 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64415.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64415.log.gz 2026-04-17T12:54:23.293 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64727.log 2026-04-17T12:54:23.293 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64694.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64694.log.gz 2026-04-17T12:54:23.304 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64743.log 2026-04-17T12:54:23.304 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64727.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64727.log.gz 2026-04-17T12:54:23.310 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64761.log 2026-04-17T12:54:23.310 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64743.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64743.log.gz 2026-04-17T12:54:23.319 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64744.log 2026-04-17T12:54:23.319 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64761.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64761.log.gz 2026-04-17T12:54:23.324 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64823.log 2026-04-17T12:54:23.325 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64744.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64744.log.gz 2026-04-17T12:54:23.333 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64774.log 2026-04-17T12:54:23.333 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64823.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64823.log.gz 2026-04-17T12:54:23.339 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.64849.log 2026-04-17T12:54:23.339 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64774.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64774.log.gz 2026-04-17T12:54:23.348 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65022.log 2026-04-17T12:54:23.348 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.64849.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.64849.log.gz 2026-04-17T12:54:23.353 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65051.log 2026-04-17T12:54:23.356 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65022.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65022.log.gz 2026-04-17T12:54:23.366 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65098.log 2026-04-17T12:54:23.366 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65051.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65051.log.gz 2026-04-17T12:54:23.372 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65197.log 2026-04-17T12:54:23.372 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65098.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65098.log.gz 2026-04-17T12:54:23.380 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65222.log 2026-04-17T12:54:23.380 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65197.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65197.log.gz 2026-04-17T12:54:23.386 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65202.log 2026-04-17T12:54:23.386 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65222.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65222.log.gz 2026-04-17T12:54:23.395 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65278.log 2026-04-17T12:54:23.395 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65202.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65202.log.gz 2026-04-17T12:54:23.396 INFO:teuthology.orchestra.run.vm02.stderr: 92.8% -- replaced with /var/log/ceph/ceph-mon.c.log.gz 2026-04-17T12:54:23.404 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65277.log 2026-04-17T12:54:23.404 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65278.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65278.log.gz 2026-04-17T12:54:23.418 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65422.log 2026-04-17T12:54:23.418 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65277.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65277.log.gz 2026-04-17T12:54:23.433 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65470.log 2026-04-17T12:54:23.433 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65422.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65422.log.gz 2026-04-17T12:54:23.448 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65498.log 2026-04-17T12:54:23.448 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65470.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65470.log.gz 2026-04-17T12:54:23.463 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65546.log 2026-04-17T12:54:23.463 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65498.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65498.log.gz 2026-04-17T12:54:23.468 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65622.log 2026-04-17T12:54:23.474 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65546.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65672.log 2026-04-17T12:54:23.474 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65622.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65622.log.gz 2026-04-17T12:54:23.477 INFO:teuthology.orchestra.run.vm02.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65546.log.gz 2026-04-17T12:54:23.479 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65722.log 2026-04-17T12:54:23.488 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65672.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65672.log.gz 2026-04-17T12:54:23.493 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65770.log 2026-04-17T12:54:23.493 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65722.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65722.log.gz 2026-04-17T12:54:23.498 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65820.log 2026-04-17T12:54:23.499 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65770.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65770.log.gz 2026-04-17T12:54:23.507 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65868.log 2026-04-17T12:54:23.507 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65820.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65820.log.gz 2026-04-17T12:54:23.513 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.65918.log 2026-04-17T12:54:23.513 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65868.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65868.log.gz 2026-04-17T12:54:23.522 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.65941.log 2026-04-17T12:54:23.522 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.65918.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.65918.log.gz 2026-04-17T12:54:23.527 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.66057.log 2026-04-17T12:54:23.529 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.65941.log: 95.0% -- replaced with /var/log/ceph/ceph-client.0.65941.log.gz 2026-04-17T12:54:23.536 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.66160.log 2026-04-17T12:54:23.536 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.66057.log: 44.6% -- replaced with /var/log/ceph/ceph-client.0.66057.log.gz 2026-04-17T12:54:23.541 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.66263.log 2026-04-17T12:54:23.551 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.66160.log: 44.9% -- replaced with /var/log/ceph/ceph-client.0.66160.log.gz 2026-04-17T12:54:23.553 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.66366.log 2026-04-17T12:54:23.553 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.66263.log: 45.6% -- replaced with /var/log/ceph/ceph-client.0.66263.log.gz 2026-04-17T12:54:23.566 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/rgw.ceph.client.0.log 2026-04-17T12:54:23.566 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.66366.log: 44.6% -- replaced with /var/log/ceph/ceph-client.0.66366.log.gz 2026-04-17T12:54:23.572 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ops-log-ceph-client.0.log 2026-04-17T12:54:23.577 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/rgw.ceph.client.0.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.0.67347.log 2026-04-17T12:54:23.582 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ops-log-ceph-client.0.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.67539.log 2026-04-17T12:54:23.591 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.67347.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.67680.log 2026-04-17T12:54:23.591 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.67539.log: 84.3% -- replaced with /var/log/ceph/ceph-client.0.67347.log.gz 2026-04-17T12:54:23.591 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.67714.log 2026-04-17T12:54:23.591 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.67680.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.67680.log.gz 2026-04-17T12:54:23.597 INFO:teuthology.orchestra.run.vm02.stderr: 83.5% -- replaced with /var/log/ceph/ceph-client.admin.67539.log.gz 2026-04-17T12:54:23.608 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.67748.log 2026-04-17T12:54:23.608 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.67714.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.67714.log.gz 2026-04-17T12:54:23.622 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.67845.log 2026-04-17T12:54:23.622 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.67748.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.67748.log.gz 2026-04-17T12:54:23.638 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.67942.log 2026-04-17T12:54:23.638 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.67845.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.67845.log.gz 2026-04-17T12:54:23.653 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68127.log 2026-04-17T12:54:23.653 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.67942.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.67942.log.gz 2026-04-17T12:54:23.667 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68161.log 2026-04-17T12:54:23.667 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68127.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68127.log.gz 2026-04-17T12:54:23.682 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68195.log 2026-04-17T12:54:23.682 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68161.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68161.log.gz 2026-04-17T12:54:23.697 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68306.log 2026-04-17T12:54:23.697 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68195.log: 82.6% -- replaced with /var/log/ceph/ceph-client.admin.68195.log.gz 2026-04-17T12:54:23.713 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68404.log 2026-04-17T12:54:23.713 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68306.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.68306.log.gz 2026-04-17T12:54:23.732 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68501.log 2026-04-17T12:54:23.741 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68404.log: 93.2% -- replaced with /var/log/ceph/ceph-client.admin.68404.log.gz 2026-04-17T12:54:23.747 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68535.log 2026-04-17T12:54:23.747 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68501.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68501.log.gz 2026-04-17T12:54:23.764 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68571.log 2026-04-17T12:54:23.764 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68535.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68535.log.gz 2026-04-17T12:54:23.783 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.68605.log 2026-04-17T12:54:23.783 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68571.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68571.log.gz 2026-04-17T12:54:23.802 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69416.log 2026-04-17T12:54:23.802 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.68605.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.68605.log.gz 2026-04-17T12:54:23.817 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69450.log 2026-04-17T12:54:23.817 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69416.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69416.log.gz 2026-04-17T12:54:23.831 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69484.log 2026-04-17T12:54:23.831 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69450.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69450.log.gz 2026-04-17T12:54:23.846 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69587.log 2026-04-17T12:54:23.846 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69484.log: 83.0% -- replaced with /var/log/ceph/ceph-client.admin.69484.log.gz 2026-04-17T12:54:23.846 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69685.log 2026-04-17T12:54:23.852 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69587.log: 84.8% -- replaced with /var/log/ceph/ceph-client.admin.69587.log.gz 2026-04-17T12:54:23.854 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69782.log 2026-04-17T12:54:23.864 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69685.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69816.log 2026-04-17T12:54:23.864 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69782.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69782.log.gz 2026-04-17T12:54:23.867 INFO:teuthology.orchestra.run.vm02.stderr: 96.7% -- replaced with /var/log/ceph/ceph-client.admin.69685.log.gz 2026-04-17T12:54:23.883 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69852.log 2026-04-17T12:54:23.883 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69816.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69816.log.gz 2026-04-17T12:54:23.888 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.69886.log 2026-04-17T12:54:23.888 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69852.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69852.log.gz 2026-04-17T12:54:23.893 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70002.log 2026-04-17T12:54:23.897 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.69886.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.69886.log.gz 2026-04-17T12:54:23.898 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70036.log 2026-04-17T12:54:23.899 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70002.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70002.log.gz 2026-04-17T12:54:23.910 INFO:teuthology.orchestra.run.vm02.stderr: 91.4% -- replaced with /var/log/ceph/ceph-mon.a.log.gz 2026-04-17T12:54:23.917 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70070.log 2026-04-17T12:54:23.917 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70036.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70036.log.gz 2026-04-17T12:54:23.925 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70173.log 2026-04-17T12:54:23.931 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70070.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70271.log 2026-04-17T12:54:23.931 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70173.log: 82.8% -- replaced with /var/log/ceph/ceph-client.admin.70070.log.gz 2026-04-17T12:54:23.932 INFO:teuthology.orchestra.run.vm02.stderr: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.70173.log.gz 2026-04-17T12:54:23.947 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70368.log 2026-04-17T12:54:23.952 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70271.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70402.log 2026-04-17T12:54:23.952 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70368.log: 92.6% -- replaced with /var/log/ceph/ceph-client.admin.70271.log.gz 2026-04-17T12:54:23.953 INFO:teuthology.orchestra.run.vm02.stderr: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70368.log.gz 2026-04-17T12:54:23.964 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70437.log 2026-04-17T12:54:23.964 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70402.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70402.log.gz 2026-04-17T12:54:23.974 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70471.log 2026-04-17T12:54:23.974 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70437.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70437.log.gz 2026-04-17T12:54:23.979 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70665.log 2026-04-17T12:54:23.981 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70471.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70471.log.gz 2026-04-17T12:54:23.992 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70699.log 2026-04-17T12:54:23.992 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70665.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70665.log.gz 2026-04-17T12:54:23.997 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70733.log 2026-04-17T12:54:23.998 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70699.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.70699.log.gz 2026-04-17T12:54:24.009 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70836.log 2026-04-17T12:54:24.009 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70733.log: 82.9% -- replaced with /var/log/ceph/ceph-client.admin.70733.log.gz 2026-04-17T12:54:24.015 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.70934.log 2026-04-17T12:54:24.017 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70836.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.70836.log.gz 2026-04-17T12:54:24.022 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71031.log 2026-04-17T12:54:24.024 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.70934.log: 89.8% -- replaced with /var/log/ceph/ceph-client.admin.70934.log.gz 2026-04-17T12:54:24.034 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71065.log 2026-04-17T12:54:24.034 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71031.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71031.log.gz 2026-04-17T12:54:24.041 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71101.log 2026-04-17T12:54:24.041 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71065.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71065.log.gz 2026-04-17T12:54:24.051 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71135.log 2026-04-17T12:54:24.051 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71101.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71101.log.gz 2026-04-17T12:54:24.057 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71259.log 2026-04-17T12:54:24.057 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71135.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71135.log.gz 2026-04-17T12:54:24.067 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71293.log 2026-04-17T12:54:24.067 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71259.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71259.log.gz 2026-04-17T12:54:24.072 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71327.log 2026-04-17T12:54:24.073 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71293.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71293.log.gz 2026-04-17T12:54:24.084 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71430.log 2026-04-17T12:54:24.084 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71327.log: 83.2% -- replaced with /var/log/ceph/ceph-client.admin.71327.log.gz 2026-04-17T12:54:24.090 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71528.log 2026-04-17T12:54:24.092 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71430.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.71430.log.gz 2026-04-17T12:54:24.102 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71625.log 2026-04-17T12:54:24.102 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71528.log: 93.3% -- replaced with /var/log/ceph/ceph-client.admin.71528.log.gz 2026-04-17T12:54:24.108 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71659.log 2026-04-17T12:54:24.108 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71625.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71625.log.gz 2026-04-17T12:54:24.118 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71695.log 2026-04-17T12:54:24.118 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71659.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71659.log.gz 2026-04-17T12:54:24.123 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.71729.log 2026-04-17T12:54:24.124 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71695.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71695.log.gz 2026-04-17T12:54:24.135 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72573.log 2026-04-17T12:54:24.135 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.71729.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.71729.log.gz 2026-04-17T12:54:24.140 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72607.log 2026-04-17T12:54:24.141 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72573.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72573.log.gz 2026-04-17T12:54:24.152 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72641.log 2026-04-17T12:54:24.152 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72607.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72607.log.gz 2026-04-17T12:54:24.157 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72744.log 2026-04-17T12:54:24.159 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72641.log: 83.0% -- replaced with /var/log/ceph/ceph-client.admin.72641.log.gz 2026-04-17T12:54:24.169 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72842.log 2026-04-17T12:54:24.169 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72744.log: 84.8% -- replaced with /var/log/ceph/ceph-client.admin.72744.log.gz 2026-04-17T12:54:24.176 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72939.log 2026-04-17T12:54:24.179 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72842.log: 90.3% -- replaced with /var/log/ceph/ceph-client.admin.72842.log.gz 2026-04-17T12:54:24.189 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.72973.log 2026-04-17T12:54:24.189 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72939.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72939.log.gz 2026-04-17T12:54:24.194 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73009.log 2026-04-17T12:54:24.196 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.72973.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.72973.log.gz 2026-04-17T12:54:24.207 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73043.log 2026-04-17T12:54:24.207 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73009.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73009.log.gz 2026-04-17T12:54:24.213 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73077.log 2026-04-17T12:54:24.213 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73043.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73043.log.gz 2026-04-17T12:54:24.223 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73174.log 2026-04-17T12:54:24.223 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73077.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.73077.log.gz 2026-04-17T12:54:24.228 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73271.log 2026-04-17T12:54:24.229 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73174.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.73174.log.gz 2026-04-17T12:54:24.239 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73650.log 2026-04-17T12:54:24.239 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73271.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.73271.log.gz 2026-04-17T12:54:24.244 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73684.log 2026-04-17T12:54:24.245 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73650.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73650.log.gz 2026-04-17T12:54:24.256 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73718.log 2026-04-17T12:54:24.256 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73684.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.73684.log.gz 2026-04-17T12:54:24.262 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73821.log 2026-04-17T12:54:24.265 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73718.log: 82.9% -- replaced with /var/log/ceph/ceph-client.admin.73718.log.gz 2026-04-17T12:54:24.275 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.73919.log 2026-04-17T12:54:24.275 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73821.log: 84.7% -- replaced with /var/log/ceph/ceph-client.admin.73821.log.gz 2026-04-17T12:54:24.281 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74016.log 2026-04-17T12:54:24.284 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.73919.log: 94.2% -- replaced with /var/log/ceph/ceph-client.admin.73919.log.gz 2026-04-17T12:54:24.294 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74050.log 2026-04-17T12:54:24.294 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74016.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74016.log.gz 2026-04-17T12:54:24.301 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74086.log 2026-04-17T12:54:24.301 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74050.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74050.log.gz 2026-04-17T12:54:24.311 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74120.log 2026-04-17T12:54:24.311 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74086.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74086.log.gz 2026-04-17T12:54:24.317 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74154.log 2026-04-17T12:54:24.317 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74120.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.74120.log.gz 2026-04-17T12:54:24.327 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74251.log 2026-04-17T12:54:24.327 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74154.log: 84.6% -- replaced with /var/log/ceph/ceph-client.admin.74154.log.gz 2026-04-17T12:54:24.333 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.74348.log 2026-04-17T12:54:24.335 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74251.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.74251.log.gz 2026-04-17T12:54:24.345 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75074.log 2026-04-17T12:54:24.345 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.74348.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.74348.log.gz 2026-04-17T12:54:24.352 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75108.log 2026-04-17T12:54:24.352 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75074.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75074.log.gz 2026-04-17T12:54:24.362 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75142.log 2026-04-17T12:54:24.362 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75108.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75108.log.gz 2026-04-17T12:54:24.368 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75245.log 2026-04-17T12:54:24.370 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75142.log: 82.9% -- replaced with /var/log/ceph/ceph-client.admin.75142.log.gz 2026-04-17T12:54:24.380 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75343.log 2026-04-17T12:54:24.380 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75245.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.75245.log.gz 2026-04-17T12:54:24.386 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75440.log 2026-04-17T12:54:24.390 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75343.log: 96.2% -- replaced with /var/log/ceph/ceph-client.admin.75343.log.gz 2026-04-17T12:54:24.401 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75474.log 2026-04-17T12:54:24.401 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75440.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75440.log.gz 2026-04-17T12:54:24.408 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75510.log 2026-04-17T12:54:24.408 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75474.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75474.log.gz 2026-04-17T12:54:24.418 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75544.log 2026-04-17T12:54:24.418 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75510.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75510.log.gz 2026-04-17T12:54:24.424 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75578.log 2026-04-17T12:54:24.424 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75544.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.75544.log.gz 2026-04-17T12:54:24.439 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75675.log 2026-04-17T12:54:24.439 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75578.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.75578.log.gz 2026-04-17T12:54:24.444 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75772.log 2026-04-17T12:54:24.446 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75675.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.75675.log.gz 2026-04-17T12:54:24.456 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.75869.log 2026-04-17T12:54:24.456 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75772.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.75772.log.gz 2026-04-17T12:54:24.461 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76150.log 2026-04-17T12:54:24.463 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.75869.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.75869.log.gz 2026-04-17T12:54:24.473 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76184.log 2026-04-17T12:54:24.473 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76150.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76150.log.gz 2026-04-17T12:54:24.479 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76218.log 2026-04-17T12:54:24.479 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76184.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76184.log.gz 2026-04-17T12:54:24.489 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76321.log 2026-04-17T12:54:24.489 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76218.log: 83.1% -- replaced with /var/log/ceph/ceph-client.admin.76218.log.gz 2026-04-17T12:54:24.495 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76419.log 2026-04-17T12:54:24.497 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76321.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.76321.log.gz 2026-04-17T12:54:24.507 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76516.log 2026-04-17T12:54:24.507 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76419.log: 89.4% -- replaced with /var/log/ceph/ceph-client.admin.76419.log.gz 2026-04-17T12:54:24.513 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76550.log 2026-04-17T12:54:24.513 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76516.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76516.log.gz 2026-04-17T12:54:24.523 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76586.log 2026-04-17T12:54:24.523 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76550.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76550.log.gz 2026-04-17T12:54:24.529 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76620.log 2026-04-17T12:54:24.529 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76586.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76586.log.gz 2026-04-17T12:54:24.539 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76654.log 2026-04-17T12:54:24.539 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76620.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.76620.log.gz 2026-04-17T12:54:24.546 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76751.log 2026-04-17T12:54:24.549 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76654.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.76654.log.gz 2026-04-17T12:54:24.559 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76848.log 2026-04-17T12:54:24.559 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76751.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.76751.log.gz 2026-04-17T12:54:24.566 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.76945.log 2026-04-17T12:54:24.568 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76848.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.76848.log.gz 2026-04-17T12:54:24.578 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77042.log 2026-04-17T12:54:24.578 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.76945.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.76945.log.gz 2026-04-17T12:54:24.583 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77139.log 2026-04-17T12:54:24.585 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77042.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.77042.log.gz 2026-04-17T12:54:24.595 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77236.log 2026-04-17T12:54:24.595 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77139.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.77139.log.gz 2026-04-17T12:54:24.600 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77333.log 2026-04-17T12:54:24.602 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77236.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.77236.log.gz 2026-04-17T12:54:24.612 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77430.log 2026-04-17T12:54:24.612 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77333.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.77333.log.gz 2026-04-17T12:54:24.618 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77527.log 2026-04-17T12:54:24.620 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77430.log: 84.9% -- replaced with /var/log/ceph/ceph-client.admin.77430.log.gz 2026-04-17T12:54:24.630 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77624.log 2026-04-17T12:54:24.630 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77527.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.77527.log.gz 2026-04-17T12:54:24.635 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77721.log 2026-04-17T12:54:24.637 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77624.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.77624.log.gz 2026-04-17T12:54:24.647 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77818.log 2026-04-17T12:54:24.647 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77721.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.77721.log.gz 2026-04-17T12:54:24.652 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.77915.log 2026-04-17T12:54:24.654 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77818.log: 85.4% -- replaced with /var/log/ceph/ceph-client.admin.77818.log.gz 2026-04-17T12:54:24.664 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78012.log 2026-04-17T12:54:24.664 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.77915.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.77915.log.gz 2026-04-17T12:54:24.669 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78109.log 2026-04-17T12:54:24.672 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78012.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.78012.log.gz 2026-04-17T12:54:24.682 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78214.log 2026-04-17T12:54:24.682 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78109.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.78109.log.gz 2026-04-17T12:54:24.688 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78311.log 2026-04-17T12:54:24.690 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78214.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.78214.log.gz 2026-04-17T12:54:24.700 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78408.log 2026-04-17T12:54:24.700 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78311.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78311.log.gz 2026-04-17T12:54:24.705 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78505.log 2026-04-17T12:54:24.707 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78408.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78408.log.gz 2026-04-17T12:54:24.717 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78602.log 2026-04-17T12:54:24.717 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78505.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.78505.log.gz 2026-04-17T12:54:24.723 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78699.log 2026-04-17T12:54:24.725 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78602.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78602.log.gz 2026-04-17T12:54:24.733 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78796.log 2026-04-17T12:54:24.733 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78699.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.78699.log.gz 2026-04-17T12:54:24.739 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78893.log 2026-04-17T12:54:24.741 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78796.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.78796.log.gz 2026-04-17T12:54:24.751 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.78990.log 2026-04-17T12:54:24.754 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78893.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78893.log.gz 2026-04-17T12:54:24.764 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79087.log 2026-04-17T12:54:24.764 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.78990.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.78990.log.gz 2026-04-17T12:54:24.770 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79184.log 2026-04-17T12:54:24.773 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79087.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.79087.log.gz 2026-04-17T12:54:24.783 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79281.log 2026-04-17T12:54:24.783 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79184.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.79184.log.gz 2026-04-17T12:54:24.789 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79378.log 2026-04-17T12:54:24.792 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79281.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.79281.log.gz 2026-04-17T12:54:24.802 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79475.log 2026-04-17T12:54:24.802 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79378.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.79378.log.gz 2026-04-17T12:54:24.808 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79572.log 2026-04-17T12:54:24.810 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79475.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.79475.log.gz 2026-04-17T12:54:24.820 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79669.log 2026-04-17T12:54:24.820 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79572.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.79572.log.gz 2026-04-17T12:54:24.826 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79766.log 2026-04-17T12:54:24.828 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79669.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.79669.log.gz 2026-04-17T12:54:24.833 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79863.log 2026-04-17T12:54:24.839 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79766.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.79960.log 2026-04-17T12:54:24.839 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79863.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.79766.log.gz 2026-04-17T12:54:24.841 INFO:teuthology.orchestra.run.vm02.stderr: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.79863.log.gz 2026-04-17T12:54:24.848 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80057.log 2026-04-17T12:54:24.850 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.79960.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.79960.log.gz 2026-04-17T12:54:24.860 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80154.log 2026-04-17T12:54:24.860 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80057.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.80057.log.gz 2026-04-17T12:54:24.866 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80251.log 2026-04-17T12:54:24.868 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80154.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.80154.log.gz 2026-04-17T12:54:24.878 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80348.log 2026-04-17T12:54:24.878 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80251.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.80251.log.gz 2026-04-17T12:54:24.884 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80445.log 2026-04-17T12:54:24.886 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80348.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.80348.log.gz 2026-04-17T12:54:24.896 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80542.log 2026-04-17T12:54:24.896 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80445.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.80445.log.gz 2026-04-17T12:54:24.902 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80639.log 2026-04-17T12:54:24.905 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80542.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.80542.log.gz 2026-04-17T12:54:24.915 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80736.log 2026-04-17T12:54:24.915 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80639.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.80639.log.gz 2026-04-17T12:54:24.922 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80833.log 2026-04-17T12:54:24.924 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80736.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.80736.log.gz 2026-04-17T12:54:24.933 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.80930.log 2026-04-17T12:54:24.933 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80833.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.80833.log.gz 2026-04-17T12:54:24.940 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81027.log 2026-04-17T12:54:24.944 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.80930.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.80930.log.gz 2026-04-17T12:54:24.954 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81124.log 2026-04-17T12:54:24.954 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81027.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.81027.log.gz 2026-04-17T12:54:24.960 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81221.log 2026-04-17T12:54:24.963 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81124.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.81124.log.gz 2026-04-17T12:54:24.973 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81318.log 2026-04-17T12:54:24.973 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81221.log: 85.2% -- replaced with /var/log/ceph/ceph-client.admin.81221.log.gz 2026-04-17T12:54:24.978 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81415.log 2026-04-17T12:54:24.980 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81318.log: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.81318.log.gz 2026-04-17T12:54:24.990 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81512.log 2026-04-17T12:54:24.990 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81415.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.81415.log.gz 2026-04-17T12:54:25.005 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81609.log 2026-04-17T12:54:25.010 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81512.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81706.log 2026-04-17T12:54:25.010 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81609.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.81512.log.gz 2026-04-17T12:54:25.013 INFO:teuthology.orchestra.run.vm02.stderr: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.81609.log.gz 2026-04-17T12:54:25.026 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81803.log 2026-04-17T12:54:25.031 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81706.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81900.log 2026-04-17T12:54:25.031 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81803.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.81706.log.gz 2026-04-17T12:54:25.032 INFO:teuthology.orchestra.run.vm02.stderr: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.81803.log.gz 2026-04-17T12:54:25.047 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.81997.log 2026-04-17T12:54:25.052 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81900.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82094.log 2026-04-17T12:54:25.052 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.81997.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.81900.log.gz 2026-04-17T12:54:25.053 INFO:teuthology.orchestra.run.vm02.stderr: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.81997.log.gz 2026-04-17T12:54:25.068 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82191.log 2026-04-17T12:54:25.073 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82094.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82288.log 2026-04-17T12:54:25.073 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82191.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82094.log.gz 2026-04-17T12:54:25.074 INFO:teuthology.orchestra.run.vm02.stderr: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82191.log.gz 2026-04-17T12:54:25.089 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82385.log 2026-04-17T12:54:25.094 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82288.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82482.log 2026-04-17T12:54:25.094 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82385.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.82288.log.gz 2026-04-17T12:54:25.095 INFO:teuthology.orchestra.run.vm02.stderr: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.82385.log.gz 2026-04-17T12:54:25.110 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82579.log 2026-04-17T12:54:25.115 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82482.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82676.log 2026-04-17T12:54:25.115 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82579.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82482.log.gz 2026-04-17T12:54:25.117 INFO:teuthology.orchestra.run.vm02.stderr: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.82579.log.gz 2026-04-17T12:54:25.131 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.82773.log 2026-04-17T12:54:25.136 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82676.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.319879.log 2026-04-17T12:54:25.136 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.82773.log: 85.3% -- replaced with /var/log/ceph/ceph-client.admin.82676.log.gz 2026-04-17T12:54:25.139 INFO:teuthology.orchestra.run.vm02.stderr: 85.0% -- replaced with /var/log/ceph/ceph-client.admin.82773.log.gz 2026-04-17T12:54:25.152 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.319913.log 2026-04-17T12:54:25.152 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.319879.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.319879.log.gz 2026-04-17T12:54:25.162 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.319947.log 2026-04-17T12:54:25.162 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.319913.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.319913.log.gz 2026-04-17T12:54:25.168 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.320050.log 2026-04-17T12:54:25.171 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.319947.log: 82.8% -- replaced with /var/log/ceph/ceph-client.admin.319947.log.gz 2026-04-17T12:54:25.181 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.320150.log 2026-04-17T12:54:25.181 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.320050.log: 85.1% -- replaced with /var/log/ceph/ceph-client.admin.320050.log.gz 2026-04-17T12:54:25.186 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.320247.log 2026-04-17T12:54:25.196 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.320150.log: gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.320281.log 2026-04-17T12:54:25.196 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.320247.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.320247.log.gz 2026-04-17T12:54:25.213 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.320317.log 2026-04-17T12:54:25.213 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.320281.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.320281.log.gz 2026-04-17T12:54:25.227 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.320351.log 2026-04-17T12:54:25.227 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.320317.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.320317.log.gz 2026-04-17T12:54:25.241 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.537215.log 2026-04-17T12:54:25.241 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.320351.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.320351.log.gz 2026-04-17T12:54:25.255 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.0.537438.log 2026-04-17T12:54:25.255 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.537215.log: 9.2% -- replaced with /var/log/ceph/ceph-client.0.537215.log.gz 2026-04-17T12:54:25.269 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/ceph-client.admin.537501.log 2026-04-17T12:54:25.269 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.0.537438.log: 8.2% -- replaced with /var/log/ceph/ceph-client.0.537438.log.gz 2026-04-17T12:54:25.283 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/ceph-client.admin.537501.log: 0.0% -- replaced with /var/log/ceph/ceph-client.admin.537501.log.gz 2026-04-17T12:54:25.596 INFO:teuthology.orchestra.run.vm02.stderr: 86.8% -- replaced with /var/log/ceph/ceph-client.admin.320150.log.gz 2026-04-17T12:54:25.654 INFO:teuthology.orchestra.run.vm02.stderr: 92.3% -- replaced with /var/log/ceph/ops-log-ceph-client.0.log.gz 2026-04-17T12:54:53.405 INFO:teuthology.orchestra.run.vm02.stderr: 2026-04-17T12:54:53.405 INFO:teuthology.orchestra.run.vm02.stderr:gzip: /var/log/ceph/ceph-osd.3.log.gz: No space left on device 2026-04-17T12:54:53.414 INFO:teuthology.orchestra.run.vm02.stderr: 2026-04-17T12:54:53.414 INFO:teuthology.orchestra.run.vm02.stderr:gzip: /var/log/ceph/ceph-osd.0.log.gz: No space left on device 2026-04-17T12:55:01.598 INFO:teuthology.orchestra.run.vm02.stderr: 91.2% -- replaced with /var/log/ceph/rgw.ceph.client.0.log.gz 2026-04-17T12:55:02.696 INFO:teuthology.orchestra.run.vm02.stderr: 93.3% -- replaced with /var/log/ceph/ceph-osd.2.log.gz 2026-04-17T12:55:17.671 INFO:teuthology.orchestra.run.vm02.stderr: 93.3% -- replaced with /var/log/ceph/ceph-osd.1.log.gz 2026-04-17T12:55:17.673 INFO:teuthology.orchestra.run.vm02.stderr: 2026-04-17T12:55:17.673 INFO:teuthology.orchestra.run.vm02.stderr:real 0m54.604s 2026-04-17T12:55:17.673 INFO:teuthology.orchestra.run.vm02.stderr:user 2m28.375s 2026-04-17T12:55:17.673 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m9.149s 2026-04-17T12:55:17.673 DEBUG:teuthology.orchestra.run:got remote process result: 123 2026-04-17T12:55:17.673 ERROR:teuthology.run_tasks:Manager failed: ceph Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2001, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 1181, in cluster yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 2011, in task ctx.managers[config['cluster']].wait_for_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2919, in wait_for_clean num_active_clean = self.get_num_active_clean() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2698, in get_num_active_clean pgs = self.get_pg_stats() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 2464, in get_pg_stats out = self.raw_cluster_cmd('pg', 'dump', '--format=json') File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1696, in raw_cluster_cmd return self.run_cluster_cmd(**kwargs).stdout.getvalue() File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph_manager.py", line 1687, in run_cluster_cmd return self.controller.run(**kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 124: 'sudo adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage timeout 120 ceph --cluster ceph pg dump --format=json' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 1996, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/ceph.py", line 263, in ceph_log run.wait( File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 123: "time sudo find /var/log/ceph -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose --" 2026-04-17T12:55:17.673 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-04-17T12:55:17.676 ERROR:teuthology.contextutil:Saw exception from nested tasks Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/teuthology/teuthology/task/install/__init__.py", line 644, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' 2026-04-17T12:55:17.676 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-04-17T12:55:17.676 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-04-17T12:55:17.715 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-04-17T12:55:17.716 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-04-17T12:55:17.750 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-04-17T12:55:17.750 DEBUG:teuthology.orchestra.run.vm02:> 2026-04-17T12:55:17.750 DEBUG:teuthology.orchestra.run.vm02:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-04-17T12:55:17.750 DEBUG:teuthology.orchestra.run.vm02:> sudo yum -y remove $d || true 2026-04-17T12:55:17.750 DEBUG:teuthology.orchestra.run.vm02:> done 2026-04-17T12:55:17.755 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-04-17T12:55:17.755 DEBUG:teuthology.orchestra.run.vm06:> 2026-04-17T12:55:17.755 DEBUG:teuthology.orchestra.run.vm06:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-04-17T12:55:17.755 DEBUG:teuthology.orchestra.run.vm06:> sudo yum -y remove $d || true 2026-04-17T12:55:17.755 DEBUG:teuthology.orchestra.run.vm06:> done 2026-04-17T12:55:17.760 INFO:teuthology.task.install.rpm:Removing packages: ceph-radosgw, ceph-test, ceph, ceph-base, cephadm, ceph-immutable-object-cache, ceph-mgr, ceph-mgr-dashboard, ceph-mgr-diskprediction-local, ceph-mgr-rook, ceph-mgr-cephadm, ceph-fuse, ceph-volume, librados-devel, libcephfs2, libcephfs-devel, librados2, librbd1, python3-rados, python3-rgw, python3-cephfs, python3-rbd, rbd-fuse, rbd-mirror, rbd-nbd on rpm system. 2026-04-17T12:55:17.760 DEBUG:teuthology.orchestra.run.vm08:> 2026-04-17T12:55:17.760 DEBUG:teuthology.orchestra.run.vm08:> for d in ceph-radosgw ceph-test ceph ceph-base cephadm ceph-immutable-object-cache ceph-mgr ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-rook ceph-mgr-cephadm ceph-fuse ceph-volume librados-devel libcephfs2 libcephfs-devel librados2 librbd1 python3-rados python3-rgw python3-cephfs python3-rbd rbd-fuse rbd-mirror rbd-nbd ; do 2026-04-17T12:55:17.760 DEBUG:teuthology.orchestra.run.vm08:> sudo yum -y remove $d || true 2026-04-17T12:55:17.760 DEBUG:teuthology.orchestra.run.vm08:> done 2026-04-17T12:55:17.958 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repo Size 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 103 M 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout: mailcap noarch 2.1.49-5.el9.0.2 @baseos 78 k 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:Remove 2 Packages 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 103 M 2026-04-17T12:55:17.959 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-04-17T12:55:17.961 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-04-17T12:55:17.961 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-04-17T12:55:17.972 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-04-17T12:55:17.972 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-04-17T12:55:17.987 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.002 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-04-17T12:55:18.024 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.024 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:18.024 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-17T12:55:18.024 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-04-17T12:55:18.024 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-04-17T12:55:18.024 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.032 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.039 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.055 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-17T12:55:18.077 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-04-17T12:55:18.077 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:18.077 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repo Size 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 103 M 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout: mailcap noarch 2.1.49-5.el9.0.2 @baseos 78 k 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:Remove 2 Packages 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 103 M 2026-04-17T12:55:18.078 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-04-17T12:55:18.082 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-04-17T12:55:18.082 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-04-17T12:55:18.096 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-04-17T12:55:18.096 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-04-17T12:55:18.117 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.118 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-17T12:55:18.118 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.127 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-04-17T12:55:18.160 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.160 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:18.160 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-radosgw@*.service" escaped as "ceph-radosgw@\x2a.service". 2026-04-17T12:55:18.160 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-radosgw.target". 2026-04-17T12:55:18.160 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-radosgw.target". 2026-04-17T12:55:18.160 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:18.165 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.176 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.191 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.201 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-04-17T12:55:18.259 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.401 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.457 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 365 M 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Remove 3 Packages 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 366 M 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-04-17T12:55:18.485 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-04-17T12:55:18.551 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.633 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-04-17T12:55:18.676 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.685 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/3 2026-04-17T12:55:18.712 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-17T12:55:18.759 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-17T12:55:18.807 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.849 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-17T12:55:18.849 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/3 2026-04-17T12:55:18.849 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-17T12:55:18.852 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-17T12:55:18.852 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 1/2 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout: ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:18.901 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-04-17T12:55:18.961 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : mailcap-2.1.49-5.el9.0.2.noarch 2/2 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout: ceph-radosgw-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout: mailcap-2.1.49-5.el9.0.2.noarch 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:18.964 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-04-17T12:55:19.125 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:19.132 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: ceph x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 0 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 6.8 M 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 19 M 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:================================================================================ 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:Remove 8 Packages 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 28 M 2026-04-17T12:55:19.134 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-04-17T12:55:19.138 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-04-17T12:55:19.138 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-04-17T12:55:19.160 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-04-17T12:55:19.160 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-04-17T12:55:19.205 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-04-17T12:55:19.210 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 365 M 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout: socat x86_64 1.7.4.1-8.el9 @appstream 1.1 M 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet x86_64 1.6.1-20.el9 @appstream 195 k 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:Remove 3 Packages 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 366 M 2026-04-17T12:55:19.211 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/8 2026-04-17T12:55:19.212 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-04-17T12:55:19.215 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-04-17T12:55:19.215 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-04-17T12:55:19.217 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-04-17T12:55:19.217 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-04-17T12:55:19.221 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-04-17T12:55:19.224 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-04-17T12:55:19.226 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-04-17T12:55:19.240 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-04-17T12:55:19.240 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-04-17T12:55:19.248 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 7/8 2026-04-17T12:55:19.249 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:19.249 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-17T12:55:19.249 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-04-17T12:55:19.249 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-04-17T12:55:19.249 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.249 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 7/8 2026-04-17T12:55:19.258 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 7/8 2026-04-17T12:55:19.288 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:19.297 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-04-17T12:55:19.305 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/3 2026-04-17T12:55:19.309 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-17T12:55:19.323 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 8/8 2026-04-17T12:55:19.323 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:19.323 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-17T12:55:19.323 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-04-17T12:55:19.323 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-04-17T12:55:19.323 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.324 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-17T12:55:19.324 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 8/8 2026-04-17T12:55:19.401 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-17T12:55:19.401 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/3 2026-04-17T12:55:19.401 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : socat-1.7.4.1-8.el9.x86_64 2/3 2026-04-17T12:55:19.415 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 8/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 3/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-04-17T12:55:19.416 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-04-17T12:55:19.446 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : xmlstarlet-1.6.1-20.el9.x86_64 3/3 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout: ceph-test-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout: socat-1.7.4.1-8.el9.x86_64 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout: xmlstarlet-1.6.1-20.el9.x86_64 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.456 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout:Removed: 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: lua-5.4.4-4.el9.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: unzip-6.0-59.el9.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: zip-3.0-35.el9.x86_64 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.468 INFO:teuthology.orchestra.run.vm08.stdout:Complete! 2026-04-17T12:55:19.592 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:19.660 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: ceph x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 0 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 6.8 M 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 19 M 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: lua x86_64 5.4.4-4.el9 @appstream 593 k 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel x86_64 5.4.4-4.el9 @crb 49 k 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: luarocks noarch 3.9.2-5.el9 @epel 692 k 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: unzip x86_64 6.0-59.el9 @baseos 389 k 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: zip x86_64 3.0-35.el9 @baseos 724 k 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:================================================================================ 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout:Remove 8 Packages 2026-04-17T12:55:19.661 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.662 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 28 M 2026-04-17T12:55:19.662 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-04-17T12:55:19.664 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-04-17T12:55:19.664 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-04-17T12:55:19.688 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-04-17T12:55:19.689 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-04-17T12:55:19.693 INFO:teuthology.orchestra.run.vm08.stdout:Dependencies resolved. 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout:==================================================================================================================== 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: Package Arch Version Repository Size 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout:==================================================================================================================== 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout:Removing: 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-base x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 24 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout:Removing dependent packages: 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 447 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 2.9 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 938 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-dashboard noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 148 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 66 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-rook noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 567 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-osd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 54 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-volume noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 1.4 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: rbd-mirror x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 11 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout:Removing unused dependencies: 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: c-ares x86_64 1.19.1-2.el9_4 @baseos 279 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 99 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-grafana-dashboards noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 990 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 60 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-mgr-modules-core noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 1.6 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-prometheus-alerts noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 57 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ceph-selinux x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 138 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: cryptsetup x86_64 2.7.2-4.el9 @baseos 722 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 @appstream 68 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 @appstream 11 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 @appstream 39 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libcephsqlite x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 409 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libgfortran x86_64 11.5.0-11.el9 @baseos 2.8 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libquadmath x86_64 11.5.0-11.el9 @baseos 330 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libradosstriper1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 792 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: libxslt x86_64 1.1.34-14.el9_7.1 @appstream 751 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: nvme-cli x86_64 2.13-1.el9 @baseos 6.8 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: protobuf x86_64 3.14.0-17.el9_7 @appstream 3.5 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 @crb 2.9 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 816 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-chardet noarch 4.0.0-5.el9 @77d52b2cce1347aa9f3fc60d8b93d222 1.4 M 2026-04-17T12:55:19.700 INFO:teuthology.orchestra.run.vm08.stdout: python3-cheroot noarch 10.0.1-5.el9 @epel 682 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-cherrypy noarch 18.10.0-5.el9 @epel 1.0 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 @baseos 4.5 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-devel x86_64 3.9.23-2.el9 @appstream 765 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-idna noarch 2.10-7.el9_4.1 @77d52b2cce1347aa9f3fc60d8b93d222 513 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 @appstream 1.1 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpatch noarch 1.21-16.el9 @0d57cd3fe20446e8b1c08da162742194 55 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-jsonpointer noarch 2.0-4.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 34 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 @appstream 30 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 @appstream 1.7 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-oauthlib noarch 3.1.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 888 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-ply noarch 3.11-14.el9.0.1 @baseos 430 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-prettytable noarch 0.7.2-27.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 166 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-protobuf noarch 3.14.0-17.el9_7 @appstream 1.4 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 @appstream 622 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 @appstream 1.0 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 @baseos 635 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pysocks noarch 1.7.1-12.el9.0.1 @77d52b2cce1347aa9f3fc60d8b93d222 88 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-pytz noarch 2021.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 176 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests noarch 2.25.1-10.el9_6 @baseos 405 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 72 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-toml noarch 0.10.2-6.el9.0.1 @appstream 99 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 @baseos 746 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: qatlib x86_64 24.09.0-1.el9 @appstream 588 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: qatlib-service x86_64 24.09.0-1.el9 @appstream 64 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: smartmontools x86_64 1:7.2-9.el9 @baseos 1.9 M 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout:Transaction Summary 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout:==================================================================================================================== 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout:Remove 111 Packages 2026-04-17T12:55:19.701 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:19.702 INFO:teuthology.orchestra.run.vm08.stdout:Freed space: 687 M 2026-04-17T12:55:19.702 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction check 2026-04-17T12:55:19.726 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-04-17T12:55:19.727 INFO:teuthology.orchestra.run.vm08.stdout:Transaction check succeeded. 2026-04-17T12:55:19.727 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction test 2026-04-17T12:55:19.732 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/8 2026-04-17T12:55:19.735 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : luarocks-3.9.2-5.el9.noarch 2/8 2026-04-17T12:55:19.735 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:19.737 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-devel-5.4.4-4.el9.x86_64 3/8 2026-04-17T12:55:19.740 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : zip-3.0-35.el9.x86_64 4/8 2026-04-17T12:55:19.743 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : unzip-6.0-59.el9.x86_64 5/8 2026-04-17T12:55:19.745 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : lua-5.4.4-4.el9.x86_64 6/8 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 7/8 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mds@*.service" escaped as "ceph-mds@\x2a.service". 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mds.target". 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mds.target". 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.766 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 7/8 2026-04-17T12:55:19.773 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 7/8 2026-04-17T12:55:19.827 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 8/8 2026-04-17T12:55:19.827 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:19.828 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mon@*.service" escaped as "ceph-mon@\x2a.service". 2026-04-17T12:55:19.828 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mon.target". 2026-04-17T12:55:19.828 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mon.target". 2026-04-17T12:55:19.828 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.828 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 8/8 2026-04-17T12:55:19.839 INFO:teuthology.orchestra.run.vm08.stdout:Transaction test succeeded. 2026-04-17T12:55:19.839 INFO:teuthology.orchestra.run.vm08.stdout:Running transaction 2026-04-17T12:55:19.863 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 8/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 1/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 3/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-5.4.4-4.el9.x86_64 4/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : lua-devel-5.4.4-4.el9.x86_64 5/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : luarocks-3.9.2-5.el9.noarch 6/8 2026-04-17T12:55:19.914 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : unzip-6.0-59.el9.x86_64 7/8 2026-04-17T12:55:19.964 INFO:teuthology.orchestra.run.vm02.stdout: Verifying : zip-3.0-35.el9.x86_64 8/8 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout:Removed: 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: ceph-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: lua-5.4.4-4.el9.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: lua-devel-5.4.4-4.el9.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: luarocks-3.9.2-5.el9.noarch 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: unzip-6.0-59.el9.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: zip-3.0-35.el9.x86_64 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:19.965 INFO:teuthology.orchestra.run.vm02.stdout:Complete! 2026-04-17T12:55:19.999 INFO:teuthology.orchestra.run.vm08.stdout: Preparing : 1/1 2026-04-17T12:55:19.999 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 1/111 2026-04-17T12:55:19.999 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.010 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 1/111 2026-04-17T12:55:20.030 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 2/111 2026-04-17T12:55:20.030 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.030 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-17T12:55:20.030 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-04-17T12:55:20.030 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-04-17T12:55:20.030 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:20.031 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 2/111 2026-04-17T12:55:20.047 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 2/111 2026-04-17T12:55:20.141 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.149 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 3/111 2026-04-17T12:55:20.149 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 4/111 2026-04-17T12:55:20.174 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 4/111 2026-04-17T12:55:20.176 INFO:teuthology.orchestra.run.vm02.stdout:Dependencies resolved. 2026-04-17T12:55:20.181 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/111 2026-04-17T12:55:20.181 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/111 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout:==================================================================================================================== 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: Package Arch Version Repository Size 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout:==================================================================================================================== 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout:Removing: 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 24 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout:Removing dependent packages: 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-immutable-object-cache x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 447 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 2.9 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 938 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 148 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-diskprediction-local noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 66 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-rook noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 567 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-osd x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 54 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-volume noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 1.4 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: rbd-mirror x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 11 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout:Removing unused dependencies: 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: abseil-cpp x86_64 20211102.0-4.el9 @epel 1.9 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: c-ares x86_64 1.19.1-2.el9_4 @baseos 279 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 99 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-grafana-dashboards noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 990 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 60 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 1.6 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-prometheus-alerts noarch 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph-noarch 57 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ceph-selinux x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 138 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: cryptsetup x86_64 2.7.2-4.el9 @baseos 722 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas x86_64 3.0.4-8.el9.0.1 @appstream 68 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-netlib x86_64 3.0.4-8.el9.0.1 @appstream 11 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: flexiblas-openblas-openmp x86_64 3.0.4-8.el9.0.1 @appstream 39 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: gperftools-libs x86_64 2.9.1-3.el9 @epel 1.4 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: grpc-data noarch 1.46.7-10.el9 @epel 13 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: ledmon-libs x86_64 1.1.0-3.el9 @baseos 80 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libcephsqlite x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 409 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libconfig x86_64 1.7.2-9.el9 @baseos 220 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libgfortran x86_64 11.5.0-11.el9 @baseos 2.8 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: liboath x86_64 2.6.12-1.el9 @epel 94 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libquadmath x86_64 11.5.0-11.el9 @baseos 330 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 792 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libstoragemgmt x86_64 1.10.1-1.el9 @appstream 685 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libunwind x86_64 1.6.2-1.el9 @epel 170 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: libxslt x86_64 1.1.34-14.el9_7.1 @appstream 751 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli x86_64 2.13-1.el9 @baseos 6.8 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: openblas x86_64 0.3.29-1.el9 @appstream 112 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: openblas-openmp x86_64 0.3.29-1.el9 @appstream 46 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: pciutils x86_64 3.7.0-7.el9 @baseos 216 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: protobuf x86_64 3.14.0-17.el9_7 @appstream 3.5 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: protobuf-compiler x86_64 3.14.0-17.el9_7 @crb 2.9 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh noarch 2.13.2-5.el9 @epel 3.9 M 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: python3-autocommand noarch 2.2.2-8.el9 @epel 82 k 2026-04-17T12:55:20.184 INFO:teuthology.orchestra.run.vm02.stdout: python3-babel noarch 2.9.1-2.el9 @appstream 27 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-backports-tarfile noarch 1.2.0-1.el9 @epel 254 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-bcrypt x86_64 3.2.2-1.el9 @epel 87 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools noarch 4.2.4-1.el9 @epel 93 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common x86_64 2:20.2.0-21.gc03ba9ecf58.el9.clyso @ceph 816 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-certifi noarch 2023.05.07-4.el9 @epel 6.3 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-cffi x86_64 1.14.5-5.el9 @baseos 1.0 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-chardet noarch 4.0.0-5.el9 @77d52b2cce1347aa9f3fc60d8b93d222 1.4 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-cheroot noarch 10.0.1-5.el9 @epel 682 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy noarch 18.10.0-5.el9 @epel 1.0 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-cryptography x86_64 36.0.1-5.el9_6 @baseos 4.5 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-devel x86_64 3.9.23-2.el9 @appstream 765 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-google-auth noarch 1:2.45.0-1.el9 @epel 1.4 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio x86_64 1.46.7-10.el9 @epel 6.7 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-grpcio-tools x86_64 1.46.7-10.el9 @epel 418 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-idna noarch 2.10-7.el9_4.1 @77d52b2cce1347aa9f3fc60d8b93d222 513 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb noarch 5.3.1-1.el9 @epel 747 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-isodate noarch 0.6.1-3.el9 @epel 203 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco noarch 8.2.1-3.el9 @epel 3.7 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-classes noarch 3.2.1-5.el9 @epel 24 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-collections noarch 3.0.0-8.el9 @epel 55 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-context noarch 6.0.1-3.el9 @epel 31 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-functools noarch 3.5.0-2.el9 @epel 33 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco-text noarch 4.0.0-2.el9 @epel 51 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jinja2 noarch 2.11.3-8.el9_5 @appstream 1.1 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpatch noarch 1.21-16.el9 @0d57cd3fe20446e8b1c08da162742194 55 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-jsonpointer noarch 2.0-4.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 34 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes noarch 1:26.1.0-3.el9 @epel 21 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-libstoragemgmt x86_64 1.10.1-1.el9 @appstream 832 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-lxml x86_64 4.6.5-3.el9 @appstream 4.2 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-markupsafe x86_64 1.1.1-12.el9 @appstream 60 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-more-itertools noarch 8.12.0-2.el9 @epel 378 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-msgpack x86_64 1.0.3-2.el9 @epel 264 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-natsort noarch 7.1.1-5.el9 @epel 215 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy x86_64 1:1.23.5-2.el9_7 @appstream 30 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-numpy-f2py x86_64 1:1.23.5-2.el9_7 @appstream 1.7 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-oauthlib noarch 3.1.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 888 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-packaging noarch 20.9-5.el9 @appstream 248 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-ply noarch 3.11-14.el9.0.1 @baseos 430 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend noarch 3.1.0-2.el9 @epel 20 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable noarch 0.7.2-27.el9.0.1 @0d57cd3fe20446e8b1c08da162742194 166 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-protobuf noarch 3.14.0-17.el9_7 @appstream 1.4 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyOpenSSL noarch 21.0.0-1.el9 @epel 389 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1 noarch 0.4.8-7.el9_7 @appstream 622 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyasn1-modules noarch 0.4.8-7.el9_7 @appstream 1.0 M 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pycparser noarch 2.20-6.el9 @baseos 745 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pyparsing noarch 2.4.7-9.el9.0.1 @baseos 635 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pysocks noarch 1.7.1-12.el9.0.1 @77d52b2cce1347aa9f3fc60d8b93d222 88 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytz noarch 2021.1-5.el9 @0d57cd3fe20446e8b1c08da162742194 176 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-repoze-lru noarch 0.7-16.el9 @epel 83 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests noarch 2.25.1-10.el9_6 @baseos 405 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib noarch 1.3.0-12.el9 @appstream 119 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes noarch 2.5.1-5.el9 @epel 459 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa noarch 4.9-2.el9 @epel 202 k 2026-04-17T12:55:20.185 INFO:teuthology.orchestra.run.vm02.stdout: python3-saml noarch 1.16.0-1.el9 @epel 730 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-scipy x86_64 1.9.3-2.el9 @appstream 72 M 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora noarch 5.0.0-2.el9 @epel 96 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-toml noarch 0.10.2-6.el9.0.1 @appstream 99 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-typing-extensions noarch 4.15.0-1.el9 @epel 447 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-urllib3 noarch 1.26.5-6.el9_7.1 @baseos 746 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket-client noarch 1.2.3-2.el9 @epel 319 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-xmlsec x86_64 1.3.13-1.el9 @epel 158 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: python3-zc-lockfile noarch 2.0-10.el9 @epel 35 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: qatlib x86_64 24.09.0-1.el9 @appstream 588 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: qatlib-service x86_64 24.09.0-1.el9 @appstream 64 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: qatzip-libs x86_64 1.3.1-1.el9 @appstream 148 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools x86_64 1:7.2-9.el9 @baseos 1.9 M 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1 x86_64 1.2.29-13.el9 @appstream 596 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: xmlsec1-openssl x86_64 1.2.29-13.el9 @appstream 281 k 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout:Transaction Summary 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout:==================================================================================================================== 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout:Remove 111 Packages 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout:Freed space: 687 M 2026-04-17T12:55:20.186 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction check 2026-04-17T12:55:20.196 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/111 2026-04-17T12:55:20.211 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cherrypy-18.10.0-5.el9.noarch 7/111 2026-04-17T12:55:20.213 INFO:teuthology.orchestra.run.vm02.stdout:Transaction check succeeded. 2026-04-17T12:55:20.213 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction test 2026-04-17T12:55:20.218 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cheroot-10.0.1-5.el9.noarch 8/111 2026-04-17T12:55:20.227 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/111 2026-04-17T12:55:20.231 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/111 2026-04-17T12:55:20.250 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 11/111 2026-04-17T12:55:20.250 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.251 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-17T12:55:20.251 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-04-17T12:55:20.251 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-04-17T12:55:20.251 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:20.256 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 11/111 2026-04-17T12:55:20.264 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 11/111 2026-04-17T12:55:20.278 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.279 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 12/111 2026-04-17T12:55:20.279 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.279 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-17T12:55:20.279 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:20.287 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 12/111 2026-04-17T12:55:20.298 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 12/111 2026-04-17T12:55:20.300 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/111 2026-04-17T12:55:20.305 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/111 2026-04-17T12:55:20.311 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jinja2-2.11.3-8.el9_5.noarch 15/111 2026-04-17T12:55:20.339 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/111 2026-04-17T12:55:20.340 INFO:teuthology.orchestra.run.vm02.stdout:Transaction test succeeded. 2026-04-17T12:55:20.340 INFO:teuthology.orchestra.run.vm02.stdout:Running transaction 2026-04-17T12:55:20.348 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 17/111 2026-04-17T12:55:20.351 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 18/111 2026-04-17T12:55:20.360 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 19/111 2026-04-17T12:55:20.372 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 20/111 2026-04-17T12:55:20.372 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 21/111 2026-04-17T12:55:20.379 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 21/111 2026-04-17T12:55:20.405 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.480 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 22/111 2026-04-17T12:55:20.516 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 23/111 2026-04-17T12:55:20.518 INFO:teuthology.orchestra.run.vm02.stdout: Preparing : 1/1 2026-04-17T12:55:20.518 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 1/111 2026-04-17T12:55:20.523 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 24/111 2026-04-17T12:55:20.526 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-rook-2:20.2.0-21.gc03ba9ecf58.el9.clyso 1/111 2026-04-17T12:55:20.527 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 25/111 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 2/111 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-mgr@*.service" escaped as "ceph-mgr@\x2a.service". 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-mgr.target". 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-mgr.target". 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:20.544 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 2/111 2026-04-17T12:55:20.546 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.550 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-17T12:55:20.551 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-04-17T12:55:20.551 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:20.551 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-17T12:55:20.559 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 2/111 2026-04-17T12:55:20.582 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-17T12:55:20.594 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 27/111 2026-04-17T12:55:20.597 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 28/111 2026-04-17T12:55:20.611 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 29/111 2026-04-17T12:55:20.617 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cryptography-36.0.1-5.el9_6.x86_64 30/111 2026-04-17T12:55:20.620 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : protobuf-compiler-3.14.0-17.el9_7.x86_64 31/111 2026-04-17T12:55:20.623 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 32/111 2026-04-17T12:55:20.642 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 33/111 2026-04-17T12:55:20.642 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.642 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-17T12:55:20.642 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-04-17T12:55:20.642 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-04-17T12:55:20.642 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:20.643 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 33/111 2026-04-17T12:55:20.652 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 33/111 2026-04-17T12:55:20.655 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 34/111 2026-04-17T12:55:20.658 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 35/111 2026-04-17T12:55:20.660 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-modules-core-2:20.2.0-21.gc03ba9ecf58.e 3/111 2026-04-17T12:55:20.660 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 4/111 2026-04-17T12:55:20.661 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 36/111 2026-04-17T12:55:20.664 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 37/111 2026-04-17T12:55:20.668 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 38/111 2026-04-17T12:55:20.671 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 39/111 2026-04-17T12:55:20.671 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 40/111 2026-04-17T12:55:20.681 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.684 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-dashboard-2:20.2.0-21.gc03ba9ecf58.el9. 4/111 2026-04-17T12:55:20.689 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-influxdb-5.3.1-1.el9.noarch 5/111 2026-04-17T12:55:20.690 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/111 2026-04-17T12:55:20.702 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-cephadm-2:20.2.0-21.gc03ba9ecf58.el9.cl 6/111 2026-04-17T12:55:20.709 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cherrypy-18.10.0-5.el9.noarch 7/111 2026-04-17T12:55:20.714 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cheroot-10.0.1-5.el9.noarch 8/111 2026-04-17T12:55:20.724 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-tools-1.46.7-10.el9.x86_64 9/111 2026-04-17T12:55:20.726 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 40/111 2026-04-17T12:55:20.728 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-grpcio-1.46.7-10.el9.x86_64 10/111 2026-04-17T12:55:20.744 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 41/111 2026-04-17T12:55:20.748 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 42/111 2026-04-17T12:55:20.752 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 11/111 2026-04-17T12:55:20.752 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.752 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-osd@*.service" escaped as "ceph-osd@\x2a.service". 2026-04-17T12:55:20.752 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-osd.target". 2026-04-17T12:55:20.752 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-osd.target". 2026-04-17T12:55:20.752 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:20.754 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 11/111 2026-04-17T12:55:20.758 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-requests-2.25.1-10.el9_6.noarch 43/111 2026-04-17T12:55:20.764 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-osd-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86_ 11/111 2026-04-17T12:55:20.768 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 44/111 2026-04-17T12:55:20.779 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 45/111 2026-04-17T12:55:20.784 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 12/111 2026-04-17T12:55:20.784 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.784 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-volume@*.service" escaped as "ceph-volume@\x2a.service". 2026-04-17T12:55:20.784 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:20.786 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 46/111 2026-04-17T12:55:20.791 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-urllib3-1.26.5-6.el9_7.1.noarch 47/111 2026-04-17T12:55:20.792 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 12/111 2026-04-17T12:55:20.797 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 48/111 2026-04-17T12:55:20.804 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-volume-2:20.2.0-21.gc03ba9ecf58.el9.clyso.n 12/111 2026-04-17T12:55:20.807 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-collections-3.0.0-8.el9.noarch 13/111 2026-04-17T12:55:20.812 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.815 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-text-4.0.0-2.el9.noarch 14/111 2026-04-17T12:55:20.823 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jinja2-2.11.3-8.el9_5.noarch 15/111 2026-04-17T12:55:20.852 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 49/111 2026-04-17T12:55:20.854 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-saml-1.16.0-1.el9.noarch 16/111 2026-04-17T12:55:20.862 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-babel-2.9.1-2.el9.noarch 17/111 2026-04-17T12:55:20.865 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-classes-3.2.1-5.el9.noarch 18/111 2026-04-17T12:55:20.870 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-numpy-1:1.23.5-2.el9_7.x86_64 50/111 2026-04-17T12:55:20.874 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyOpenSSL-21.0.0-1.el9.noarch 19/111 2026-04-17T12:55:20.876 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 51/111 2026-04-17T12:55:20.881 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 52/111 2026-04-17T12:55:20.882 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-asyncssh-2.13.2-5.el9.noarch 20/111 2026-04-17T12:55:20.882 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 21/111 2026-04-17T12:55:20.883 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 53/111 2026-04-17T12:55:20.887 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : libgfortran-11.5.0-11.el9.x86_64 54/111 2026-04-17T12:55:20.890 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-diskprediction-local-2:20.2.0-21.gc03ba 21/111 2026-04-17T12:55:20.890 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 55/111 2026-04-17T12:55:20.918 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 56/111 2026-04-17T12:55:20.918 INFO:teuthology.orchestra.run.vm08.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:20.918 INFO:teuthology.orchestra.run.vm08.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-17T12:55:20.918 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:20.919 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 56/111 2026-04-17T12:55:20.931 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 56/111 2026-04-17T12:55:20.933 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 57/111 2026-04-17T12:55:20.936 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : flexiblas-3.0.4-8.el9.0.1.x86_64 58/111 2026-04-17T12:55:20.939 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ply-3.11-14.el9.0.1.noarch 59/111 2026-04-17T12:55:20.942 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-idna-2.10-7.el9_4.1.noarch 60/111 2026-04-17T12:55:20.947 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pysocks-1.7.1-12.el9.0.1.noarch 61/111 2026-04-17T12:55:20.951 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:20.952 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyasn1-0.4.8-7.el9_7.noarch 62/111 2026-04-17T12:55:20.959 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 63/111 2026-04-17T12:55:20.967 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 64/111 2026-04-17T12:55:20.973 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 65/111 2026-04-17T12:55:20.977 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 66/111 2026-04-17T12:55:20.979 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 67/111 2026-04-17T12:55:20.982 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 68/111 2026-04-17T12:55:20.985 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 69/111 2026-04-17T12:55:20.988 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 70/111 2026-04-17T12:55:20.990 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpatch-1.21-16.el9.noarch 22/111 2026-04-17T12:55:20.991 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-toml-0.10.2-6.el9.0.1.noarch 71/111 2026-04-17T12:55:20.995 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 72/111 2026-04-17T12:55:20.998 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 73/111 2026-04-17T12:55:21.008 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/111 2026-04-17T12:55:21.009 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-scipy-1.9.3-2.el9.x86_64 23/111 2026-04-17T12:55:21.012 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-devel-3.9.23-2.el9.x86_64 75/111 2026-04-17T12:55:21.015 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-jsonpointer-2.0-4.el9.0.1.noarch 76/111 2026-04-17T12:55:21.016 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-xmlsec-1.3.13-1.el9.x86_64 24/111 2026-04-17T12:55:21.020 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/111 2026-04-17T12:55:21.021 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-lxml-4.6.5-3.el9.x86_64 25/111 2026-04-17T12:55:21.024 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 78/111 2026-04-17T12:55:21.027 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 79/111 2026-04-17T12:55:21.033 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 80/111 2026-04-17T12:55:21.037 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-17T12:55:21.037 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/libstoragemgmt.service". 2026-04-17T12:55:21.037 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:21.037 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-protobuf-3.14.0-17.el9_7.noarch 81/111 2026-04-17T12:55:21.038 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-17T12:55:21.041 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 82/111 2026-04-17T12:55:21.045 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 83/111 2026-04-17T12:55:21.047 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 84/111 2026-04-17T12:55:21.049 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 85/111 2026-04-17T12:55:21.072 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 86/111 2026-04-17T12:55:21.072 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-04-17T12:55:21.072 INFO:teuthology.orchestra.run.vm08.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-04-17T12:55:21.072 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:21.073 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: libstoragemgmt-1.10.1-1.el9.x86_64 26/111 2026-04-17T12:55:21.079 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-f2py-1:1.23.5-2.el9_7.x86_64 27/111 2026-04-17T12:55:21.080 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 86/111 2026-04-17T12:55:21.080 INFO:teuthology.orchestra.run.vm08.stdout:warning: file /etc/logrotate.d/ceph: remove failed: No such file or directory 2026-04-17T12:55:21.080 INFO:teuthology.orchestra.run.vm08.stdout: 2026-04-17T12:55:21.082 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlsec1-openssl-1.2.29-13.el9.x86_64 28/111 2026-04-17T12:55:21.087 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:21.095 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : xmlsec1-1.2.29-13.el9.x86_64 29/111 2026-04-17T12:55:21.102 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cryptography-36.0.1-5.el9_6.x86_64 30/111 2026-04-17T12:55:21.107 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : protobuf-compiler-3.14.0-17.el9_7.x86_64 31/111 2026-04-17T12:55:21.110 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-bcrypt-3.2.2-1.el9.x86_64 32/111 2026-04-17T12:55:21.111 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 86/111 2026-04-17T12:55:21.111 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 87/111 2026-04-17T12:55:21.128 INFO:teuthology.orchestra.run.vm08.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 87/111 2026-04-17T12:55:21.132 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 33/111 2026-04-17T12:55:21.132 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:21.132 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-rbd-mirror@*.service" escaped as "ceph-rbd-mirror@\x2a.service". 2026-04-17T12:55:21.132 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph-rbd-mirror.target". 2026-04-17T12:55:21.132 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-rbd-mirror.target". 2026-04-17T12:55:21.132 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:21.133 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 33/111 2026-04-17T12:55:21.133 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 88/111 2026-04-17T12:55:21.137 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 89/111 2026-04-17T12:55:21.139 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : python3-prettytable-0.7.2-27.el9.0.1.noarch 90/111 2026-04-17T12:55:21.139 INFO:teuthology.orchestra.run.vm08.stdout: Erasing : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 91/111 2026-04-17T12:55:21.142 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: rbd-mirror-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x8 33/111 2026-04-17T12:55:21.145 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-context-6.0.1-3.el9.noarch 34/111 2026-04-17T12:55:21.149 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-packaging-20.9-5.el9.noarch 35/111 2026-04-17T12:55:21.152 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-portend-3.1.0-2.el9.noarch 36/111 2026-04-17T12:55:21.155 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-tempora-5.0.0-2.el9.noarch 37/111 2026-04-17T12:55:21.158 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-functools-3.5.0-2.el9.noarch 38/111 2026-04-17T12:55:21.162 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-routes-2.5.1-5.el9.noarch 39/111 2026-04-17T12:55:21.162 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 40/111 2026-04-17T12:55:21.215 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:21.218 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-mgr-k8sevents-2:20.2.0-21.gc03ba9ecf58.el9. 40/111 2026-04-17T12:55:21.229 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-kubernetes-1:26.1.0-3.el9.noarch 41/111 2026-04-17T12:55:21.234 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-oauthlib-1.3.0-12.el9.noarch 42/111 2026-04-17T12:55:21.243 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-requests-2.25.1-10.el9_6.noarch 43/111 2026-04-17T12:55:21.249 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-google-auth-1:2.45.0-1.el9.noarch 44/111 2026-04-17T12:55:21.260 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-rsa-4.9-2.el9.noarch 45/111 2026-04-17T12:55:21.267 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-modules-0.4.8-7.el9_7.noarch 46/111 2026-04-17T12:55:21.272 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-urllib3-1.26.5-6.el9_7.1.noarch 47/111 2026-04-17T12:55:21.277 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cffi-1.14.5-5.el9.x86_64 48/111 2026-04-17T12:55:21.326 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pycparser-2.20-6.el9.noarch 49/111 2026-04-17T12:55:21.338 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-numpy-1:1.23.5-2.el9_7.x86_64 50/111 2026-04-17T12:55:21.341 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-netlib-3.0.4-8.el9.0.1.x86_64 51/111 2026-04-17T12:55:21.342 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:21.343 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-openblas-openmp-3.0.4-8.el9.0.1.x86_64 52/111 2026-04-17T12:55:21.346 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-openmp-0.3.29-1.el9.x86_64 53/111 2026-04-17T12:55:21.350 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : libgfortran-11.5.0-11.el9.x86_64 54/111 2026-04-17T12:55:21.354 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-libstoragemgmt-1.10.1-1.el9.x86_64 55/111 2026-04-17T12:55:21.364 DEBUG:teuthology.orchestra.run.vm06:> sudo yum clean all 2026-04-17T12:55:21.374 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 56/111 2026-04-17T12:55:21.375 INFO:teuthology.orchestra.run.vm02.stdout:Glob pattern passed to enable, but globs are not supported for this. 2026-04-17T12:55:21.375 INFO:teuthology.orchestra.run.vm02.stdout:Invalid unit name "ceph-immutable-object-cache@*.service" escaped as "ceph-immutable-object-cache@\x2a.service". 2026-04-17T12:55:21.375 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:21.375 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 56/111 2026-04-17T12:55:21.384 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-immutable-object-cache-2:20.2.0-21.gc03ba9e 56/111 2026-04-17T12:55:21.386 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : openblas-0.3.29-1.el9.x86_64 57/111 2026-04-17T12:55:21.389 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : flexiblas-3.0.4-8.el9.0.1.x86_64 58/111 2026-04-17T12:55:21.391 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ply-3.11-14.el9.0.1.noarch 59/111 2026-04-17T12:55:21.394 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-idna-2.10-7.el9_4.1.noarch 60/111 2026-04-17T12:55:21.400 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pysocks-1.7.1-12.el9.0.1.noarch 61/111 2026-04-17T12:55:21.404 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyasn1-0.4.8-7.el9_7.noarch 62/111 2026-04-17T12:55:21.410 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-cachetools-4.2.4-1.el9.noarch 63/111 2026-04-17T12:55:21.418 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-chardet-4.0.0-5.el9.noarch 64/111 2026-04-17T12:55:21.424 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-oauthlib-3.1.1-5.el9.noarch 65/111 2026-04-17T12:55:21.427 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-websocket-client-1.2.3-2.el9.noarch 66/111 2026-04-17T12:55:21.430 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-certifi-2023.05.07-4.el9.noarch 67/111 2026-04-17T12:55:21.432 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-repoze-lru-0.7-16.el9.noarch 68/111 2026-04-17T12:55:21.435 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jaraco-8.2.1-3.el9.noarch 69/111 2026-04-17T12:55:21.437 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-more-itertools-8.12.0-2.el9.noarch 70/111 2026-04-17T12:55:21.440 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-toml-0.10.2-6.el9.0.1.noarch 71/111 2026-04-17T12:55:21.443 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pytz-2021.1-5.el9.noarch 72/111 2026-04-17T12:55:21.446 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-pyparsing-2.4.7-9.el9.0.1.noarch 73/111 2026-04-17T12:55:21.454 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-backports-tarfile-1.2.0-1.el9.noarch 74/111 2026-04-17T12:55:21.459 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-devel-3.9.23-2.el9.x86_64 75/111 2026-04-17T12:55:21.461 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-jsonpointer-2.0-4.el9.0.1.noarch 76/111 2026-04-17T12:55:21.465 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-typing-extensions-4.15.0-1.el9.noarch 77/111 2026-04-17T12:55:21.468 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-isodate-0.6.1-3.el9.noarch 78/111 2026-04-17T12:55:21.471 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-autocommand-2.2.2-8.el9.noarch 79/111 2026-04-17T12:55:21.476 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : grpc-data-1.46.7-10.el9.noarch 80/111 2026-04-17T12:55:21.480 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-protobuf-3.14.0-17.el9_7.noarch 81/111 2026-04-17T12:55:21.483 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-zc-lockfile-2.0-10.el9.noarch 82/111 2026-04-17T12:55:21.486 INFO:teuthology.orchestra.run.vm06.stderr:[Errno 28] No space left on device: '/var/cache/dnf/metadata_lock.pid' 2026-04-17T12:55:21.486 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-natsort-7.1.1-5.el9.noarch 83/111 2026-04-17T12:55:21.487 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-grafana-dashboards-2:20.2.0-21.gc03ba9ecf58 84/111 2026-04-17T12:55:21.489 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-prometheus-alerts-2:20.2.0-21.gc03ba9ecf58. 85/111 2026-04-17T12:55:21.508 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:55:21.508 ERROR:teuthology.run_tasks:Manager failed: install Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/task/install/__init__.py", line 220, in install yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 32, in nested yield vars File "/home/teuthos/teuthology/teuthology/task/install/__init__.py", line 644, in task yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/teuthology/teuthology/task/install/__init__.py", line 640, in task with contextutil.nested( File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/teuthology/teuthology/task/install/__init__.py", line 222, in install remove_packages(ctx, config, package_list) File "/home/teuthos/teuthology/teuthology/task/install/__init__.py", line 103, in remove_packages with parallel() as p: File "/home/teuthos/teuthology/teuthology/parallel.py", line 84, in __exit__ for result in self: File "/home/teuthos/teuthology/teuthology/parallel.py", line 98, in __next__ resurrect_traceback(result) File "/home/teuthos/teuthology/teuthology/parallel.py", line 30, in resurrect_traceback raise exc.exc_info[1] File "/home/teuthos/teuthology/teuthology/parallel.py", line 23, in capture_traceback return func(*args, **kwargs) File "/home/teuthos/teuthology/teuthology/task/install/rpm.py", line 43, in _remove remote.run(args='sudo yum clean all') File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'sudo yum clean all' 2026-04-17T12:55:21.508 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-04-17T12:55:21.511 INFO:teuthology.task.clock:Checking final clock skew... 2026-04-17T12:55:21.511 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-17T12:55:21.511 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 86/111 2026-04-17T12:55:21.511 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/multi-user.target.wants/ceph.target". 2026-04-17T12:55:21.511 INFO:teuthology.orchestra.run.vm02.stdout:Removed "/etc/systemd/system/ceph.target.wants/ceph-crash.service". 2026-04-17T12:55:21.511 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:21.512 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-17T12:55:21.514 DEBUG:teuthology.orchestra.run.vm08:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-04-17T12:55:21.517 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 86/111 2026-04-17T12:55:21.518 INFO:teuthology.orchestra.run.vm02.stdout:warning: file /etc/logrotate.d/ceph: remove failed: No such file or directory 2026-04-17T12:55:21.518 INFO:teuthology.orchestra.run.vm02.stdout: 2026-04-17T12:55:21.529 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-04-17T12:55:21.529 INFO:teuthology.orchestra.run.vm02.stderr:bash: line 1: ntpq: command not found 2026-04-17T12:55:21.531 INFO:teuthology.orchestra.run.vm08.stderr:bash: line 1: ntpq: command not found 2026-04-17T12:55:21.532 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-17T12:55:21.532 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-04-17T12:55:21.532 INFO:teuthology.orchestra.run.vm06.stdout:^* mail.bmwiedemann.de 2 7 377 96 -556us[ -582us] +/- 15ms 2026-04-17T12:55:21.532 INFO:teuthology.orchestra.run.vm06.stdout:^+ mail.morbitzer.de 2 6 377 31 +1493us[+1493us] +/- 20ms 2026-04-17T12:55:21.532 INFO:teuthology.orchestra.run.vm06.stdout:^- x1.ncomputers.org 2 8 377 28 +604us[ +604us] +/- 50ms 2026-04-17T12:55:21.532 INFO:teuthology.orchestra.run.vm06.stdout:^+ ntp5.kernfusion.at 3 8 377 97 -1568us[-1595us] +/- 33ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm02.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm02.stdout:=============================================================================== 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm02.stdout:^* mail.bmwiedemann.de 2 6 377 31 -528us[ -533us] +/- 16ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm02.stdout:^+ mail.morbitzer.de 2 7 377 97 +1512us[+1458us] +/- 20ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm02.stdout:^- x1.ncomputers.org 2 8 377 95 +150us[ +147us] +/- 49ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm02.stdout:^+ ntp5.kernfusion.at 3 8 377 102 -1511us[-1565us] +/- 33ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm08.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm08.stdout:=============================================================================== 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm08.stdout:^+ ntp5.kernfusion.at 3 8 377 96 -1488us[-1527us] +/- 33ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm08.stdout:^* mail.bmwiedemann.de 2 7 377 96 -544us[ -583us] +/- 15ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm08.stdout:^+ mail.morbitzer.de 2 7 377 98 +1517us[+1478us] +/- 20ms 2026-04-17T12:55:21.534 INFO:teuthology.orchestra.run.vm08.stdout:^- x1.ncomputers.org 2 7 377 97 +690us[ +651us] +/- 48ms 2026-04-17T12:55:21.535 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-04-17T12:55:21.537 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-04-17T12:55:21.538 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-04-17T12:55:21.540 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-04-17T12:55:21.542 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-04-17T12:55:21.545 INFO:teuthology.task.internal:Duration was 2360.259277 seconds 2026-04-17T12:55:21.545 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-04-17T12:55:21.547 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-04-17T12:55:21.547 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-04-17T12:55:21.548 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-base-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x86 86/111 2026-04-17T12:55:21.548 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 87/111 2026-04-17T12:55:21.549 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-04-17T12:55:21.575 DEBUG:teuthology.orchestra.run.vm08:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-04-17T12:55:21.585 INFO:teuthology.orchestra.run.vm02.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-17T12:55:21.611 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-17T12:55:21.623 INFO:teuthology.orchestra.run.vm08.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-04-17T12:55:21.891 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-04-17T12:55:21.891 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-04-17T12:55:21.892 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-04-17T12:55:21.913 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-04-17T12:55:21.913 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-04-17T12:55:21.935 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm08.local 2026-04-17T12:55:21.935 DEBUG:teuthology.orchestra.run.vm08:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-04-17T12:55:21.963 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-04-17T12:55:21.963 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-04-17T12:55:21.964 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-04-17T12:55:21.977 DEBUG:teuthology.orchestra.run.vm08:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-04-17T12:55:21.991 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: /home/ubuntu/cephtest/archive/syslog/journalctl.log: No space left on device 2026-04-17T12:55:22.195 INFO:teuthology.orchestra.run.vm02.stdout: Running scriptlet: ceph-common-2:20.2.0-21.gc03ba9ecf58.el9.clyso.x 87/111 2026-04-17T12:55:22.201 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : qatzip-libs-1.3.1-1.el9.x86_64 88/111 2026-04-17T12:55:22.204 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-ceph-common-2:20.2.0-21.gc03ba9ecf58.el9 89/111 2026-04-17T12:55:22.206 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : python3-prettytable-0.7.2-27.el9.0.1.noarch 90/111 2026-04-17T12:55:22.206 INFO:teuthology.orchestra.run.vm02.stdout: Erasing : ceph-selinux-2:20.2.0-21.gc03ba9ecf58.el9.clyso. 91/111 2026-04-17T12:55:22.341 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:55:22.341 ERROR:teuthology.run_tasks:Manager failed: internal.syslog Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/task/internal/syslog.py", line 76, in syslog yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/teuthology/teuthology/task/internal/syslog.py", line 163, in syslog run.wait( File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm06 with status 1: 'sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log' 2026-04-17T12:55:22.341 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-04-17T12:55:22.344 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-04-17T12:55:22.344 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-04-17T12:55:22.371 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-04-17T12:55:22.397 DEBUG:teuthology.orchestra.run.vm08:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-04-17T12:55:22.431 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-04-17T12:55:22.434 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-04-17T12:55:22.435 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-04-17T12:55:22.439 DEBUG:teuthology.orchestra.run.vm08:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-04-17T12:55:22.461 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-04-17T12:55:22.464 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-04-17T12:55:22.494 INFO:teuthology.orchestra.run.vm08.stdout:kernel.core_pattern = core 2026-04-17T12:55:22.509 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-04-17T12:55:22.524 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:55:22.524 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-04-17T12:55:22.541 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:55:22.541 DEBUG:teuthology.orchestra.run.vm08:> test -e /home/ubuntu/cephtest/archive/coredump 2026-04-17T12:55:22.565 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:55:22.565 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-04-17T12:55:22.574 INFO:teuthology.task.internal:Transferring archived files... 2026-04-17T12:55:22.574 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/remote/vm02 2026-04-17T12:55:22.574 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-04-17T12:55:22.771 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/remote/vm06 2026-04-17T12:55:22.772 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-04-17T12:55:22.798 DEBUG:teuthology.misc:Transferring archived files from vm08:/home/ubuntu/cephtest/archive to /archive/supriti-2026-04-17_12:11:56-rgw-wip-sse-s3-on-v20.2.0-none-default-vps/5589/remote/vm08 2026-04-17T12:55:22.798 DEBUG:teuthology.orchestra.run.vm08:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-04-17T12:55:22.938 INFO:teuthology.task.internal:Removing archive directory... 2026-04-17T12:55:22.938 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-04-17T12:55:22.940 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-04-17T12:55:22.941 DEBUG:teuthology.orchestra.run.vm08:> rm -rf -- /home/ubuntu/cephtest/archive 2026-04-17T12:55:22.996 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-04-17T12:55:22.999 INFO:teuthology.task.internal:Not uploading archives. 2026-04-17T12:55:22.999 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-04-17T12:55:23.002 INFO:teuthology.task.internal:Tidying up after the test... 2026-04-17T12:55:23.002 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-04-17T12:55:23.004 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-04-17T12:55:23.005 DEBUG:teuthology.orchestra.run.vm08:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-04-17T12:55:23.017 INFO:teuthology.orchestra.run.vm02.stdout: 58724742 0 drwxr-xr-x 3 ubuntu ubuntu 23 Apr 17 12:55 /home/ubuntu/cephtest 2026-04-17T12:55:23.017 INFO:teuthology.orchestra.run.vm02.stdout: 75512447 0 drwxr-xr-x 2 ubuntu ubuntu 6 Apr 17 12:19 /home/ubuntu/cephtest/ceph.data 2026-04-17T12:55:23.017 INFO:teuthology.orchestra.run.vm02.stderr:rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty 2026-04-17T12:55:23.020 INFO:teuthology.orchestra.run.vm06.stdout: 58721292 0 drwxr-xr-x 3 ubuntu ubuntu 76 Apr 17 12:55 /home/ubuntu/cephtest 2026-04-17T12:55:23.020 INFO:teuthology.orchestra.run.vm06.stdout: 75515903 0 drwxr-xr-x 2 ubuntu ubuntu 6 Apr 17 12:19 /home/ubuntu/cephtest/ceph.data 2026-04-17T12:55:23.020 INFO:teuthology.orchestra.run.vm06.stdout: 58720616 4 -rw-r--r-- 1 ceph root 20 Apr 17 12:20 /home/ubuntu/cephtest/url_file 2026-04-17T12:55:23.020 INFO:teuthology.orchestra.run.vm06.stdout: 58720617 0 srwxr-xr-x 1 root root 0 Apr 17 12:20 /home/ubuntu/cephtest/rgw.opslog.ceph.client.1.sock 2026-04-17T12:55:23.021 INFO:teuthology.orchestra.run.vm06.stderr:rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty 2026-04-17T12:55:23.038 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-04-17T12:55:23.038 ERROR:teuthology.run_tasks:Manager failed: internal.base Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/task/internal/__init__.py", line 48, in base yield File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 552, in task with contextutil.nested(*subtasks): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/teuthology/teuthology/contextutil.py", line 54, in nested raise exc[1] File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 364, in create_pools yield File "/home/teuthos/teuthology/teuthology/contextutil.py", line 46, in nested if exit(*exc): File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 142, in __exit__ next(self.gen) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/rgw.py", line 269, in start_rgw rgwadmin(ctx, client, cmd=['gc', 'process', '--include-all'], check_status=True) File "/home/teuthos/src/git.local_ceph_909b66e106532fd1f1a49171c3f2eb7a193a6d0b/qa/tasks/util/rgw.py", line 34, in rgwadmin proc = remote.run( File "/home/teuthos/teuthology/teuthology/orchestra/remote.py", line 575, in run r = self._runner(client=self.ssh, name=self.shortname, **kwargs) File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 461, in run r.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/teuthos/teuthology/teuthology/run_tasks.py", line 160, in run_tasks suppress = manager.__exit__(*exc_info) File "/home/teuthos/.local/share/uv/python/cpython-3.10.19-linux-x86_64-gnu/lib/python3.10/contextlib.py", line 153, in __exit__ self.gen.throw(typ, value, traceback) File "/home/teuthos/teuthology/teuthology/task/internal/__init__.py", line 53, in base run.wait( File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 485, in wait proc.wait() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 161, in wait self._raise_for_status() File "/home/teuthos/teuthology/teuthology/orchestra/run.py", line 181, in _raise_for_status raise CommandFailedError( teuthology.exceptions.CommandFailedError: Command failed on vm02 with status 1: 'find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest' 2026-04-17T12:55:23.038 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-04-17T12:55:23.041 DEBUG:teuthology.run_tasks:Exception was not quenched, exiting: CommandFailedError: Command failed on vm02 with status 1: 'adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin --log-to-stderr --format json -n client.0 --cluster ceph gc process --include-all' 2026-04-17T12:55:23.042 INFO:teuthology.run:Summary data: description: rgw/dedup/{beast bluestore-bitmap fixed-3-rgw ignore-pg-availability overrides supported-distros/{rocky_latest} tasks/{0-install test_dedup}} duration: 2360.2592766284943 failure_reason: 'Command failed on vm02 with status 1: ''adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage radosgw-admin -n client.0 user rm --uid foo.client.0 --purge-data --cluster ceph''' flavor: default owner: supriti sentry_event: null status: fail success: false 2026-04-17T12:55:23.042 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-04-17T12:55:23.052 INFO:teuthology.orchestra.run.vm08.stdout: 58724742 0 drwxr-xr-x 3 ubuntu ubuntu 95 Apr 17 12:55 /home/ubuntu/cephtest 2026-04-17T12:55:23.052 INFO:teuthology.orchestra.run.vm08.stdout: 75512383 0 drwxr-xr-x 2 ubuntu ubuntu 6 Apr 17 12:19 /home/ubuntu/cephtest/ceph.data 2026-04-17T12:55:23.052 INFO:teuthology.orchestra.run.vm08.stdout: 58725145 4 -rw-r--r-- 1 ubuntu ubuntu 409 Apr 17 12:19 /home/ubuntu/cephtest/ceph.monmap 2026-04-17T12:55:23.052 INFO:teuthology.orchestra.run.vm08.stdout: 58725146 4 -rw-r--r-- 1 ceph root 20 Apr 17 12:20 /home/ubuntu/cephtest/url_file 2026-04-17T12:55:23.052 INFO:teuthology.orchestra.run.vm08.stdout: 58725147 0 srwxr-xr-x 1 root root 0 Apr 17 12:20 /home/ubuntu/cephtest/rgw.opslog.ceph.client.2.sock 2026-04-17T12:55:23.053 INFO:teuthology.orchestra.run.vm08.stderr:rmdir: failed to remove '/home/ubuntu/cephtest': Directory not empty 2026-04-17T12:55:23.066 INFO:teuthology.run:FAIL